[ 447.821017] env[61991]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61991) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 447.821459] env[61991]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61991) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 447.821675] env[61991]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61991) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 447.822100] env[61991]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 447.915668] env[61991]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61991) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 447.927190] env[61991]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.011s {{(pid=61991) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 448.527692] env[61991]: INFO nova.virt.driver [None req-9117e29b-289b-42ba-ab03-cc39ea29964c None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 448.596706] env[61991]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 448.596917] env[61991]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 448.597012] env[61991]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61991) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 451.638487] env[61991]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-5dfa7dd6-3515-4278-b0ef-bca13a1ef3b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.654718] env[61991]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61991) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 451.654909] env[61991]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-6d3ef06f-6972-4081-a07f-4d762a71d805 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.690044] env[61991]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 9ed02. [ 451.690223] env[61991]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.093s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 451.690743] env[61991]: INFO nova.virt.vmwareapi.driver [None req-9117e29b-289b-42ba-ab03-cc39ea29964c None None] VMware vCenter version: 7.0.3 [ 451.694173] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3616ff62-9e80-4c8c-876b-a01fb3ff454a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.715850] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e98500e-76be-4258-b965-21fc84a828c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.721576] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce42764-1587-46f6-92e3-618805f26938 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.728008] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b33dc8-8fbd-4376-9cd4-f3cde2579a5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.740651] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5386c9-aca9-48a9-96be-916ca5826bf7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.746366] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753f478e-92fd-4a5a-bea3-f81fd31ce37b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.775471] env[61991]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-ec920987-0f5d-4f23-a06f-4fa219425b3e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.780194] env[61991]: DEBUG nova.virt.vmwareapi.driver [None req-9117e29b-289b-42ba-ab03-cc39ea29964c None None] Extension org.openstack.compute already exists. {{(pid=61991) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 451.782918] env[61991]: INFO nova.compute.provider_config [None req-9117e29b-289b-42ba-ab03-cc39ea29964c None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 452.286606] env[61991]: DEBUG nova.context [None req-9117e29b-289b-42ba-ab03-cc39ea29964c None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),3d7fb67f-0b36-406d-88d5-a38bbc37cabb(cell1) {{(pid=61991) load_cells /opt/stack/nova/nova/context.py:464}} [ 452.288889] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 452.289164] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 452.290173] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 452.290626] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Acquiring lock "3d7fb67f-0b36-406d-88d5-a38bbc37cabb" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 452.290975] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Lock "3d7fb67f-0b36-406d-88d5-a38bbc37cabb" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 452.292373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Lock "3d7fb67f-0b36-406d-88d5-a38bbc37cabb" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 452.313056] env[61991]: INFO dbcounter [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Registered counter for database nova_cell0 [ 452.321484] env[61991]: INFO dbcounter [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Registered counter for database nova_cell1 [ 452.325054] env[61991]: DEBUG oslo_db.sqlalchemy.engines [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61991) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 452.325662] env[61991]: DEBUG oslo_db.sqlalchemy.engines [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61991) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 452.330079] env[61991]: ERROR nova.db.main.api [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 452.330079] env[61991]: result = function(*args, **kwargs) [ 452.330079] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 452.330079] env[61991]: return func(*args, **kwargs) [ 452.330079] env[61991]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 452.330079] env[61991]: result = fn(*args, **kwargs) [ 452.330079] env[61991]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 452.330079] env[61991]: return f(*args, **kwargs) [ 452.330079] env[61991]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 452.330079] env[61991]: return db.service_get_minimum_version(context, binaries) [ 452.330079] env[61991]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 452.330079] env[61991]: _check_db_access() [ 452.330079] env[61991]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 452.330079] env[61991]: stacktrace = ''.join(traceback.format_stack()) [ 452.330079] env[61991]: [ 452.331612] env[61991]: ERROR nova.db.main.api [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 452.331612] env[61991]: result = function(*args, **kwargs) [ 452.331612] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 452.331612] env[61991]: return func(*args, **kwargs) [ 452.331612] env[61991]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 452.331612] env[61991]: result = fn(*args, **kwargs) [ 452.331612] env[61991]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 452.331612] env[61991]: return f(*args, **kwargs) [ 452.331612] env[61991]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 452.331612] env[61991]: return db.service_get_minimum_version(context, binaries) [ 452.331612] env[61991]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 452.331612] env[61991]: _check_db_access() [ 452.331612] env[61991]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 452.331612] env[61991]: stacktrace = ''.join(traceback.format_stack()) [ 452.331612] env[61991]: [ 452.332214] env[61991]: WARNING nova.objects.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 452.332214] env[61991]: WARNING nova.objects.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Failed to get minimum service version for cell 3d7fb67f-0b36-406d-88d5-a38bbc37cabb [ 452.332611] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Acquiring lock "singleton_lock" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 452.332800] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Acquired lock "singleton_lock" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 452.333077] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Releasing lock "singleton_lock" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 452.333415] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Full set of CONF: {{(pid=61991) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 452.333560] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ******************************************************************************** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 452.333688] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Configuration options gathered from: {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 452.333825] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 452.334032] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 452.334166] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ================================================================================ {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 452.334381] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] allow_resize_to_same_host = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.334552] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] arq_binding_timeout = 300 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.334683] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] backdoor_port = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.334811] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] backdoor_socket = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.334976] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] block_device_allocate_retries = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.335153] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] block_device_allocate_retries_interval = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.335320] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cert = self.pem {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.335487] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.335653] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute_monitors = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.335821] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] config_dir = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.335990] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] config_drive_format = iso9660 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.336141] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.336310] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] config_source = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.336477] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] console_host = devstack {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.336757] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] control_exchange = nova {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.336930] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cpu_allocation_ratio = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.337102] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] daemon = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.337277] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] debug = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.337433] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] default_access_ip_network_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.337621] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] default_availability_zone = nova {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.337750] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] default_ephemeral_format = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.337909] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] default_green_pool_size = 1000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.338164] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.338332] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] default_schedule_zone = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.338532] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] disk_allocation_ratio = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.338649] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] enable_new_services = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.338828] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] enabled_apis = ['osapi_compute'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.338991] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] enabled_ssl_apis = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.339165] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] flat_injected = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.339327] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] force_config_drive = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.339485] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] force_raw_images = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.339692] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] graceful_shutdown_timeout = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.339868] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] heal_instance_info_cache_interval = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.340095] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] host = cpu-1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.340272] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.340439] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.340599] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.340813] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.340984] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_build_timeout = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.341162] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_delete_interval = 300 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.341331] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_format = [instance: %(uuid)s] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.341498] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_name_template = instance-%08x {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.341685] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_usage_audit = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.341875] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_usage_audit_period = month {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.342058] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.342233] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.342399] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] internal_service_availability_zone = internal {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.342557] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] key = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.342722] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] live_migration_retry_count = 30 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.342916] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_color = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.343097] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_config_append = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.343269] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.343429] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_dir = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.343586] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.343715] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_options = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.343877] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_rotate_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344150] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_rotate_interval_type = days {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344215] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] log_rotation_type = none {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344340] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344466] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344629] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344796] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.344951] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.345137] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] long_rpc_timeout = 1800 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.345299] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] max_concurrent_builds = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.345462] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] max_concurrent_live_migrations = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.345624] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] max_concurrent_snapshots = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.345785] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] max_local_block_devices = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.345945] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] max_logfile_count = 30 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.346116] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] max_logfile_size_mb = 200 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.346276] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] maximum_instance_delete_attempts = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.346445] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metadata_listen = 0.0.0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.346686] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metadata_listen_port = 8775 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.346891] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metadata_workers = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.347077] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] migrate_max_retries = -1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.347249] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] mkisofs_cmd = genisoimage {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.347454] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.347586] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] my_ip = 10.180.1.21 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.347749] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] network_allocate_retries = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.347928] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.348109] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.348275] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] osapi_compute_listen_port = 8774 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.348441] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] osapi_compute_unique_server_name_scope = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.348645] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] osapi_compute_workers = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.348773] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] password_length = 12 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.348932] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] periodic_enable = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.349103] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] periodic_fuzzy_delay = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.349274] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] pointer_model = usbtablet {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.349440] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] preallocate_images = none {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.349628] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] publish_errors = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.349774] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] pybasedir = /opt/stack/nova {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.349940] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ram_allocation_ratio = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.350115] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] rate_limit_burst = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.350285] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] rate_limit_except_level = CRITICAL {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.350444] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] rate_limit_interval = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.350603] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reboot_timeout = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.350762] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reclaim_instance_interval = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.350918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] record = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.351098] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reimage_timeout_per_gb = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.351269] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] report_interval = 120 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.351429] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] rescue_timeout = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.351591] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reserved_host_cpus = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.351804] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reserved_host_disk_mb = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.351979] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reserved_host_memory_mb = 512 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.352156] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] reserved_huge_pages = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.352320] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] resize_confirm_window = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.352481] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] resize_fs_using_block_device = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.352640] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] resume_guests_state_on_host_boot = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.352831] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.353015] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] rpc_response_timeout = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.353188] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] run_external_periodic_tasks = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.353352] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] running_deleted_instance_action = reap {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.353512] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.353672] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] running_deleted_instance_timeout = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.353833] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler_instance_sync_interval = 120 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354020] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_down_time = 720 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354183] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] servicegroup_driver = db {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354339] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] shell_completion = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354497] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] shelved_offload_time = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354654] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] shelved_poll_interval = 3600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354820] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] shutdown_timeout = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.354978] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] source_is_ipv6 = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.355149] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ssl_only = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.355395] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.355562] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] sync_power_state_interval = 600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.355722] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] sync_power_state_pool_size = 1000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.355888] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] syslog_log_facility = LOG_USER {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.356141] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] tempdir = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.356225] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] timeout_nbd = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.356375] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] transport_url = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.356538] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] update_resources_interval = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.356699] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_cow_images = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.356980] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_eventlog = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.357172] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_journal = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.357337] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_json = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.357497] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_rootwrap_daemon = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.357658] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_stderr = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.357818] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] use_syslog = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.357974] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vcpu_pin_set = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.358158] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plugging_is_fatal = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.358329] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plugging_timeout = 300 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.358494] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] virt_mkfs = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.358659] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] volume_usage_poll_interval = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.358824] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] watch_log_file = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.358989] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] web = /usr/share/spice-html5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 452.359196] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.359366] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.359537] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.359742] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_concurrency.disable_process_locking = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.360320] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.360520] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.360696] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.360876] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.361061] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.361235] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.361422] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.auth_strategy = keystone {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.361592] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.compute_link_prefix = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.361803] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.361989] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.dhcp_domain = novalocal {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.362201] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.enable_instance_password = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.362396] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.glance_link_prefix = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.362567] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.362744] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.363070] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.instance_list_per_project_cells = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.363267] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.list_records_by_skipping_down_cells = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.363436] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.local_metadata_per_cell = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.363608] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.max_limit = 1000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.363814] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.metadata_cache_expiration = 15 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.364000] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.neutron_default_tenant_id = default {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.364188] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.response_validation = warn {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.364361] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.use_neutron_default_nets = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.364546] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.364711] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.364883] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.365067] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.365244] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_dynamic_targets = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.365410] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_jsonfile_path = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.365593] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.365790] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.backend = dogpile.cache.memcached {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.365961] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.backend_argument = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.366147] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.config_prefix = cache.oslo {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.366320] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.dead_timeout = 60.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.366489] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.debug_cache_backend = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.366654] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.enable_retry_client = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.366819] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.enable_socket_keepalive = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.366991] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.enabled = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.367264] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.enforce_fips_mode = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.367442] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.expiration_time = 600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.367612] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.hashclient_retry_attempts = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.367783] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.367953] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_dead_retry = 300 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.368204] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_password = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.368380] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.368548] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.368713] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_pool_maxsize = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.368889] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.369069] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_sasl_enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.369255] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.369425] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.369619] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.memcache_username = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.369806] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.proxies = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.369982] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_db = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.370162] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_password = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.370339] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.370518] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.370689] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_server = localhost:6379 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.370858] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_socket_timeout = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.371032] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.redis_username = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.371205] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.retry_attempts = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.371374] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.retry_delay = 0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.371540] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.socket_keepalive_count = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.371733] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.socket_keepalive_idle = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.371915] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.socket_keepalive_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.372090] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.tls_allowed_ciphers = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.372254] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.tls_cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.372414] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.tls_certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.372579] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.tls_enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.372740] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cache.tls_keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.372913] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.373100] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.auth_type = password {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.373265] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.373444] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.373608] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.373772] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.373938] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.cross_az_attach = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.374158] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.debug = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.374332] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.endpoint_template = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.374500] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.http_retries = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.374664] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.374827] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.375009] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.os_region_name = RegionOne {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.375186] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.375351] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cinder.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.375524] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.375703] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.cpu_dedicated_set = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.375888] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.cpu_shared_set = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.376071] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.image_type_exclude_list = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.376313] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.376503] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.376673] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.376843] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.377024] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.377276] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.resource_provider_association_refresh = 300 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.377465] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.377637] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.shutdown_retry_interval = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.377826] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.378023] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] conductor.workers = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.378203] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] console.allowed_origins = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.378368] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] console.ssl_ciphers = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.378543] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] console.ssl_minimum_version = default {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.378715] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] consoleauth.enforce_session_timeout = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.378888] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] consoleauth.token_ttl = 600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.379070] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.379232] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.379397] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.379571] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.379762] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.379929] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.380104] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.380269] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.380426] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.380583] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.380741] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.region_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.380900] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.381067] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.381241] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.service_type = accelerator {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.381404] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.381563] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.381745] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.381918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.382115] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.382279] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] cyborg.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.382459] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.backend = sqlalchemy {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.382630] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.connection = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.382797] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.connection_debug = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.382967] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.connection_parameters = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.383149] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.connection_recycle_time = 3600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.383315] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.connection_trace = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.383477] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.db_inc_retry_interval = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.383642] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.db_max_retries = 20 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.383808] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.db_max_retry_interval = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.383983] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.db_retry_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.384179] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.max_overflow = 50 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.384347] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.max_pool_size = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.384512] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.max_retries = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.384685] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.384850] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.mysql_wsrep_sync_wait = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.385014] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.pool_timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.385186] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.retry_interval = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.385345] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.slave_connection = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.385509] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.sqlite_synchronous = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.385675] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] database.use_db_reconnect = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.385853] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.backend = sqlalchemy {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.386056] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.connection = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.386254] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.connection_debug = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.386430] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.connection_parameters = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.386595] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.connection_recycle_time = 3600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.386763] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.connection_trace = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.386943] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.db_inc_retry_interval = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.387307] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.db_max_retries = 20 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.387534] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.db_max_retry_interval = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.387726] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.db_retry_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.387929] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.max_overflow = 50 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.388117] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.max_pool_size = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.388289] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.max_retries = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.388463] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.388626] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.388788] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.pool_timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.388957] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.retry_interval = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.389132] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.slave_connection = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.389299] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] api_database.sqlite_synchronous = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.389478] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] devices.enabled_mdev_types = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.389658] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.389837] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.390014] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ephemeral_storage_encryption.enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.390188] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.390359] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.api_servers = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.390525] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.390688] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.390858] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.391031] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.391198] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.391362] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.debug = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.391531] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.default_trusted_certificate_ids = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.391731] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.enable_certificate_validation = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.391909] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.enable_rbd_download = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.392085] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.392259] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.392425] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.392587] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.392745] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.392912] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.num_retries = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.393095] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.rbd_ceph_conf = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.393264] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.rbd_connect_timeout = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.393435] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.rbd_pool = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.393605] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.rbd_user = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.393763] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.region_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.393927] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.394120] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.394274] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.service_type = image {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.394437] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.394606] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.394769] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.394929] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.395123] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.395294] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.verify_glance_signatures = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.395456] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] glance.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.395624] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] guestfs.debug = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.395791] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] mks.enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.396165] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.396361] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] image_cache.manager_interval = 2400 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.396532] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] image_cache.precache_concurrency = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.396704] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] image_cache.remove_unused_base_images = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.396875] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.397052] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.397233] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] image_cache.subdirectory_name = _base {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.397485] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.api_max_retries = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.397675] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.api_retry_interval = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.397869] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.398053] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.auth_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.398218] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.398378] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.398546] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.398707] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.conductor_group = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.398868] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.399035] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.399202] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.399363] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.399521] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.399734] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.399917] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.400099] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.peer_list = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.400267] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.region_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.400424] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.400593] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.serial_console_state_timeout = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.400744] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.400916] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.service_type = baremetal {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.401083] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.shard = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.401250] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.401410] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.401571] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.401758] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.401954] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.402133] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ironic.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.402318] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.402494] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] key_manager.fixed_key = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.402694] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.402877] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.barbican_api_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.403050] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.barbican_endpoint = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.403231] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.barbican_endpoint_type = public {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.403394] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.barbican_region_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.403556] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.403721] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.403927] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.404111] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.404273] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.404440] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.number_of_retries = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.404604] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.retry_delay = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.404767] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.send_service_user_token = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.404931] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.405101] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.405264] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.verify_ssl = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.405423] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican.verify_ssl_path = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.405591] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.405755] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.auth_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.405914] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.406083] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.406250] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.406411] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.406570] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.406734] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.406894] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] barbican_service_user.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.407070] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.approle_role_id = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.407236] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.approle_secret_id = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.407406] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.kv_mountpoint = secret {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.407658] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.kv_path = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.407841] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.kv_version = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.408014] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.namespace = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.408182] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.root_token_id = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.408344] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.ssl_ca_crt_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.408511] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.timeout = 60.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.408678] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.use_ssl = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.408853] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.409041] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.409210] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.auth_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.409372] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.409530] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.409717] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.409895] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.410072] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.410233] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.410394] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.410551] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.410709] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.410868] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.411037] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.region_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.411201] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.411359] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.411527] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.service_type = identity {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.411709] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.411881] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.412057] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.412220] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.412401] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.412563] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] keystone.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.412764] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.connection_uri = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.412928] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_mode = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.413115] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.413303] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_models = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.413477] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_power_governor_high = performance {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.413647] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.413812] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_power_management = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.413982] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.414162] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.device_detach_attempts = 8 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.414326] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.device_detach_timeout = 20 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.414492] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.disk_cachemodes = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.414651] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.disk_prefix = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.414817] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.enabled_perf_events = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.414980] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.file_backed_memory = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.415159] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.gid_maps = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.415318] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.hw_disk_discard = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.415474] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.hw_machine_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.415646] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_rbd_ceph_conf = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.415815] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.415978] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.416158] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_rbd_glance_store_name = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.416327] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_rbd_pool = rbd {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.416497] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_type = default {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.416659] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.images_volume_group = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.416825] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.inject_key = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.416988] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.inject_partition = -2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.417165] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.inject_password = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.417331] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.iscsi_iface = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.417493] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.iser_use_multipath = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.417674] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.417838] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418009] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_downtime = 500 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418176] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418341] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418504] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_inbound_addr = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418666] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418836] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.418998] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_scheme = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.419188] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_timeout_action = abort {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.419352] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_tunnelled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.419513] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_uri = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.419692] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.live_migration_with_native_tls = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.419868] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.max_queues = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.420043] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.420280] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.420443] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.nfs_mount_options = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.420733] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.420911] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.421088] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.421255] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.421421] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.421583] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.num_pcie_ports = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.421825] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.422040] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.pmem_namespaces = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.422212] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.quobyte_client_cfg = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.422501] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.422677] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.422845] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.423023] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.423184] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rbd_secret_uuid = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.423345] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rbd_user = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.423510] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.423683] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.423845] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rescue_image_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.424013] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rescue_kernel_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.424179] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rescue_ramdisk_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.424348] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.424507] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.rx_queue_size = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.424673] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.smbfs_mount_options = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.425006] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.425207] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.snapshot_compression = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.425377] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.snapshot_image_format = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.425638] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.425832] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.sparse_logical_volumes = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.426034] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.swtpm_enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.426918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.swtpm_group = tss {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.426918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.swtpm_user = tss {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.426918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.sysinfo_serial = unique {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.426918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.tb_cache_size = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.426918] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.tx_queue_size = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.427153] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.uid_maps = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.427255] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.use_virtio_for_bridges = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.427447] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.virt_type = kvm {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.427639] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.volume_clear = zero {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.427815] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.volume_clear_size = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.427984] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.volume_use_multipath = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.428159] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_cache_path = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.428335] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.428507] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.428673] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.428846] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.429143] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.429328] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.vzstorage_mount_user = stack {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.429499] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.429696] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.429888] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.auth_type = password {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.430062] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.430228] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.430396] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.430557] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.430716] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.430893] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.default_floating_pool = public {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.431068] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.431238] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.extension_sync_interval = 600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.431403] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.http_retries = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.431568] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.431754] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.431927] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.432116] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.432283] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.432458] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.ovs_bridge = br-int {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.432631] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.physnets = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.432831] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.region_name = RegionOne {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.432999] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.433187] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.service_metadata_proxy = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.433350] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.433519] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.service_type = network {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.433685] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.433879] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.434055] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.434219] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.434406] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.434569] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] neutron.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.434744] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] notifications.bdms_in_notifications = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.434927] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] notifications.default_level = INFO {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.435119] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] notifications.notification_format = unversioned {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.435290] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] notifications.notify_on_state_change = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.435469] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.435646] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] pci.alias = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.435817] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] pci.device_spec = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.435982] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] pci.report_in_placement = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.436171] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.436346] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.auth_type = password {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.436512] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.436674] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.436833] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.436998] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.437171] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.437331] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.437489] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.default_domain_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.437646] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.default_domain_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.437804] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.domain_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.437957] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.domain_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.438126] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.438291] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.438449] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.438609] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.438766] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.438935] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.password = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.439107] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.project_domain_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.439323] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.project_domain_name = Default {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.439533] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.project_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.439736] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.project_name = service {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.439924] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.region_name = RegionOne {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.440110] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.440274] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.440442] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.service_type = placement {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.440605] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.440767] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.440929] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.441098] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.system_scope = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.441258] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.441417] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.trust_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.441573] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.user_domain_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.441766] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.user_domain_name = Default {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.441933] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.user_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.442122] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.username = nova {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.442304] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.442465] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] placement.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.442648] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.cores = 20 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.442844] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.count_usage_from_placement = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.443031] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.443216] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.injected_file_content_bytes = 10240 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.443386] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.injected_file_path_length = 255 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.443554] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.injected_files = 5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.443720] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.instances = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.443886] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.key_pairs = 100 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.444062] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.metadata_items = 128 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.444232] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.ram = 51200 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.444393] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.recheck_quota = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.444562] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.server_group_members = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.444729] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] quota.server_groups = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.444911] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.445085] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.445263] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.image_metadata_prefilter = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.445429] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.445594] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.max_attempts = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.445770] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.max_placement_results = 1000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.445964] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.446145] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.446311] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.446484] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] scheduler.workers = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.446657] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.446827] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.447012] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.447193] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.447359] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.447530] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.447691] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.447885] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.448068] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.host_subset_size = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.448236] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.448397] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.448560] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.448724] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.isolated_hosts = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.448889] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.isolated_images = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.449063] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.449231] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.449393] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.449555] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.pci_in_placement = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.449742] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.449921] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.450097] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.450263] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.450426] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.450588] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.450749] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.track_instance_changes = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.450927] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.451118] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metrics.required = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.451286] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metrics.weight_multiplier = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.451449] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.451610] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] metrics.weight_setting = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.451958] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.452158] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] serial_console.enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.452337] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] serial_console.port_range = 10000:20000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.452511] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.452681] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.452851] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] serial_console.serialproxy_port = 6083 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.453037] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.453216] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.auth_type = password {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.453375] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.453533] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.453694] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.453857] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.454015] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.454191] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.send_service_user_token = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.454356] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.454527] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] service_user.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.454696] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.agent_enabled = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.454859] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.455192] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.455386] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.455556] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.html5proxy_port = 6082 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.455716] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.image_compression = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.455879] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.jpeg_compression = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.456050] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.playback_compression = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.456219] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.require_secure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.456387] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.server_listen = 127.0.0.1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.456556] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.456715] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.streaming_mode = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.456873] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] spice.zlib_compression = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.457045] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] upgrade_levels.baseapi = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.457220] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] upgrade_levels.compute = auto {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.457377] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] upgrade_levels.conductor = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.457533] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] upgrade_levels.scheduler = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.457697] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.457904] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.458121] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.458292] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.458457] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.458621] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.458782] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.458944] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.459119] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vendordata_dynamic_auth.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.459299] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.api_retry_count = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.459461] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.ca_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.459637] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.459801] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.cluster_name = testcl1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.459970] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.connection_pool_size = 10 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.460144] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.console_delay_seconds = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.460314] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.datastore_regex = ^datastore.* {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.460522] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.460695] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.host_password = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.460862] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.host_port = 443 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.461039] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.host_username = administrator@vsphere.local {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.461217] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.insecure = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.461378] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.integration_bridge = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.461541] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.maximum_objects = 100 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.461720] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.pbm_default_policy = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.461892] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.pbm_enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.462061] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.pbm_wsdl_location = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.462236] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.462397] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.serial_port_proxy_uri = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.462553] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.serial_port_service_uri = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.462721] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.task_poll_interval = 0.5 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.462892] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.use_linked_clone = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.463071] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.vnc_keymap = en-us {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.463238] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.vnc_port = 5900 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.463403] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vmware.vnc_port_total = 10000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.463588] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.auth_schemes = ['none'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.463763] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.464079] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.464268] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.464440] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.novncproxy_port = 6080 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.464617] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.server_listen = 127.0.0.1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.464788] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.464950] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.vencrypt_ca_certs = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.465122] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.vencrypt_client_cert = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.465284] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vnc.vencrypt_client_key = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.465466] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.465632] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.disable_deep_image_inspection = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.465797] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.465957] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.466134] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.466298] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.disable_rootwrap = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.466460] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.enable_numa_live_migration = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.466622] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.466786] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.466954] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.467129] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.libvirt_disable_apic = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.467291] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.467458] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.467651] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.467872] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.468074] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.468248] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.468415] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.468577] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.468741] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.468909] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.469108] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.469281] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.client_socket_timeout = 900 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.469447] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.default_pool_size = 1000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.469614] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.keep_alive = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.469783] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.max_header_line = 16384 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.469946] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.470122] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.ssl_ca_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.470287] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.ssl_cert_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.470447] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.ssl_key_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.470611] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.tcp_keepidle = 600 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.470795] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.470972] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] zvm.ca_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.471146] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] zvm.cloud_connector_url = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.471433] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.471606] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] zvm.reachable_timeout = 300 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.471812] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.enforce_new_defaults = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.472237] env[61991]: WARNING oslo_config.cfg [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 452.472428] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.enforce_scope = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.472610] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.policy_default_rule = default {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.472795] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.472978] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.policy_file = policy.yaml {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.473172] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.473337] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.473501] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.473660] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.473823] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.473994] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.474183] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.474361] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.connection_string = messaging:// {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.474528] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.enabled = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.474698] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.es_doc_type = notification {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.474865] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.es_scroll_size = 10000 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.475044] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.es_scroll_time = 2m {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.475213] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.filter_error_trace = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.475383] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.hmac_keys = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.475551] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.sentinel_service_name = mymaster {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.475718] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.socket_timeout = 0.1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.475883] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.trace_requests = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.476063] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler.trace_sqlalchemy = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.476244] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler_jaeger.process_tags = {} {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.476406] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler_jaeger.service_name_prefix = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.476570] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] profiler_otlp.service_name_prefix = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.476736] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] remote_debug.host = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.476895] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] remote_debug.port = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.477083] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.477251] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.477415] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.477579] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.477743] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.477907] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.478078] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.478244] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.478411] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.478582] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.478742] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.478914] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.479094] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.479268] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.479440] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.479609] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.479804] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.479991] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.480182] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.480350] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.480517] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.480685] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.480852] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.481030] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.481197] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.481359] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.481524] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.481705] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.481888] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.482075] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.ssl = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.482252] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.482426] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.482589] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.482788] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.482974] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.483174] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.483370] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.483543] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_notifications.retry = -1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.483729] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.483909] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.484095] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.auth_section = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.484265] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.auth_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.484427] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.cafile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.484586] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.certfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.484750] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.collect_timing = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.484911] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.connect_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.485079] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.connect_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.485244] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.endpoint_id = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.485404] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.endpoint_override = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.485568] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.insecure = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.485726] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.keyfile = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.485887] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.max_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.486057] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.min_version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.486219] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.region_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.486380] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.retriable_status_codes = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.486541] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.service_name = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.486698] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.service_type = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.486863] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.split_loggers = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.487031] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.status_code_retries = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.487195] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.status_code_retry_delay = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.487355] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.timeout = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.487515] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.valid_interfaces = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.487674] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_limit.version = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.487840] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_reports.file_event_handler = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.488010] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.488178] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] oslo_reports.log_dir = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.488349] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.488510] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.488669] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.488835] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.489008] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.489173] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.489344] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.489503] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_ovs_privileged.group = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.489682] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.489890] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.490077] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.490265] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] vif_plug_ovs_privileged.user = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.490412] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.490601] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.490778] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.490952] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.491138] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.491308] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.491475] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.491645] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.491851] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.492044] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.isolate_vif = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.492222] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.492393] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.492565] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.492743] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.492908] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] os_vif_ovs.per_port_bridge = False {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.493091] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] privsep_osbrick.capabilities = [21] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.493255] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] privsep_osbrick.group = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.493414] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] privsep_osbrick.helper_command = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.493581] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.493771] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.493936] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] privsep_osbrick.user = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.494128] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.494291] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] nova_sys_admin.group = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.494450] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] nova_sys_admin.helper_command = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.494617] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.494782] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.494941] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] nova_sys_admin.user = None {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 452.495086] env[61991]: DEBUG oslo_service.service [None req-3da27dcd-2b6b-4f35-bffd-a1b0c3d4d303 None None] ******************************************************************************** {{(pid=61991) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 452.495574] env[61991]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 452.998819] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Getting list of instances from cluster (obj){ [ 452.998819] env[61991]: value = "domain-c8" [ 452.998819] env[61991]: _type = "ClusterComputeResource" [ 452.998819] env[61991]: } {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 453.000028] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ac521f-34d1-469d-961e-8535ef265f88 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.008855] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Got total of 0 instances {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 453.009400] env[61991]: WARNING nova.virt.vmwareapi.driver [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 453.009875] env[61991]: INFO nova.virt.node [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Generated node identity d748992a-e0bf-4ec2-9c17-0e373360e5a3 [ 453.010121] env[61991]: INFO nova.virt.node [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Wrote node identity d748992a-e0bf-4ec2-9c17-0e373360e5a3 to /opt/stack/data/n-cpu-1/compute_id [ 453.512455] env[61991]: WARNING nova.compute.manager [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Compute nodes ['d748992a-e0bf-4ec2-9c17-0e373360e5a3'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 454.518238] env[61991]: INFO nova.compute.manager [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 455.523447] env[61991]: WARNING nova.compute.manager [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 455.523812] env[61991]: DEBUG oslo_concurrency.lockutils [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 455.523933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 455.524105] env[61991]: DEBUG oslo_concurrency.lockutils [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 455.524275] env[61991]: DEBUG nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 455.525179] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b51263c-3dba-4bad-b670-38fe64a926a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.533356] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41305cd7-b5ac-4819-99a6-4f4bbeb04113 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.546669] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ed7853-68d3-48a0-9f35-717d70bd3150 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.553353] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6bb37cd-879a-4507-9447-369f1295ad69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.582334] env[61991]: DEBUG nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181535MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 455.582470] env[61991]: DEBUG oslo_concurrency.lockutils [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 455.582681] env[61991]: DEBUG oslo_concurrency.lockutils [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 456.084941] env[61991]: WARNING nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] No compute node record for cpu-1:d748992a-e0bf-4ec2-9c17-0e373360e5a3: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host d748992a-e0bf-4ec2-9c17-0e373360e5a3 could not be found. [ 456.588388] env[61991]: INFO nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: d748992a-e0bf-4ec2-9c17-0e373360e5a3 [ 458.096211] env[61991]: DEBUG nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 458.096583] env[61991]: DEBUG nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 458.242889] env[61991]: INFO nova.scheduler.client.report [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] [req-23a33995-ed23-4255-9c59-469f0fd949b9] Created resource provider record via placement API for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 458.261318] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717d9b2c-4570-4423-88aa-7cc00fc61c25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 458.269129] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98112ea6-e02f-47f4-a9ab-974ac95c114e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 458.298632] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb93275-9d9e-455c-a0f4-1899192b7242 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 458.305983] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8adbf7-ecaa-40e6-b3b3-4cafc8e76239 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 458.319446] env[61991]: DEBUG nova.compute.provider_tree [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 458.857984] env[61991]: DEBUG nova.scheduler.client.report [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 458.858248] env[61991]: DEBUG nova.compute.provider_tree [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 0 to 1 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 458.858500] env[61991]: DEBUG nova.compute.provider_tree [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 458.915049] env[61991]: DEBUG nova.compute.provider_tree [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 1 to 2 during operation: update_traits {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 459.419846] env[61991]: DEBUG nova.compute.resource_tracker [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 459.420249] env[61991]: DEBUG oslo_concurrency.lockutils [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.837s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 459.420249] env[61991]: DEBUG nova.service [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Creating RPC server for service compute {{(pid=61991) start /opt/stack/nova/nova/service.py:186}} [ 459.434271] env[61991]: DEBUG nova.service [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] Join ServiceGroup membership for this service compute {{(pid=61991) start /opt/stack/nova/nova/service.py:203}} [ 459.434517] env[61991]: DEBUG nova.servicegroup.drivers.db [None req-924fc94a-dd88-41c3-b763-f7c943f40142 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61991) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 470.437888] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 470.941024] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Getting list of instances from cluster (obj){ [ 470.941024] env[61991]: value = "domain-c8" [ 470.941024] env[61991]: _type = "ClusterComputeResource" [ 470.941024] env[61991]: } {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 470.942272] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c17819-00d0-46b1-9f2b-0be238d8e30b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.950764] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Got total of 0 instances {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 470.950984] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 470.951288] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Getting list of instances from cluster (obj){ [ 470.951288] env[61991]: value = "domain-c8" [ 470.951288] env[61991]: _type = "ClusterComputeResource" [ 470.951288] env[61991]: } {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 470.952126] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20bf964-c37f-4ea7-990d-e1fcc07eb291 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.958935] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Got total of 0 instances {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 495.373318] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.373602] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.878057] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 496.421140] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.421430] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.423886] env[61991]: INFO nova.compute.claims [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 496.479799] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "d978d68a-0709-4a32-bbaf-3bfa006be85c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.480155] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "d978d68a-0709-4a32-bbaf-3bfa006be85c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.560519] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "f32fee48-de95-4ade-9ed0-092786d3f39c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.560519] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "f32fee48-de95-4ade-9ed0-092786d3f39c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.984835] env[61991]: DEBUG nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 497.062362] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 497.268254] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquiring lock "ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.269711] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Lock "ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.406555] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.406672] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.515061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.572298] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffc3de9-36e5-4da0-8112-2602cccd4e49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.583687] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fcf53f-9a95-4ace-a454-606dfd308c84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.617573] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.621019] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92cae5f-1abf-4f86-a054-e93718e3ae85 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.626457] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071aabdd-0c23-419d-85dc-cd1896528686 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.640843] env[61991]: DEBUG nova.compute.provider_tree [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 497.774723] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 497.794315] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquiring lock "cdecebd4-5075-414d-bce5-7bee66c68cba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.795041] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Lock "cdecebd4-5075-414d-bce5-7bee66c68cba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.909131] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 498.144833] env[61991]: DEBUG nova.scheduler.client.report [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 498.297567] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 498.312570] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.436024] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.651241] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.652334] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 498.660016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.143s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.660016] env[61991]: INFO nova.compute.claims [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 498.826286] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.172483] env[61991]: DEBUG nova.compute.utils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 499.174147] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 499.174419] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 499.570549] env[61991]: DEBUG nova.policy [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a385cf261940a19eb17046958920ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3015571d285418aae0fca50b563394e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 499.683024] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 499.828199] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c965a5-a831-4c10-af7b-853ef760b919 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.837496] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c78a5a1-0fa8-4e9a-bb86-cbb6f989b2ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.867269] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a96c299-9e87-4bf5-8131-5305a4714e2f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.874824] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c355855b-178a-411d-bc35-615dd865388f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.888134] env[61991]: DEBUG nova.compute.provider_tree [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 500.396327] env[61991]: DEBUG nova.scheduler.client.report [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 500.591256] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Successfully created port: d093b2e7-784c-4581-bc31-9c6be567a256 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 500.691738] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 500.726541] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 500.726872] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 500.727079] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 500.727311] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 500.727479] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 500.727648] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 500.727914] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 500.728176] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 500.728653] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 500.728856] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 500.729041] env[61991]: DEBUG nova.virt.hardware [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 500.730016] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8af681a-15ce-484e-8889-0c6e744b2594 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.738136] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4643423-91fe-4d42-9b70-c9f4e3f63c72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.753492] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dbc2bd-2039-4332-b471-e33efe8ce199 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.901828] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.244s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.902486] env[61991]: DEBUG nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 500.908027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.291s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.909651] env[61991]: INFO nova.compute.claims [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 501.417790] env[61991]: DEBUG nova.compute.utils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 501.425019] env[61991]: DEBUG nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 501.926035] env[61991]: DEBUG nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 502.069726] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa71309-e980-4b59-9c2c-3699eecc69c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.081576] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93929f0b-99b1-4dcd-abe7-ba0b2d4d387d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.121973] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec124418-0502-4d79-bfe4-ab0126720bd8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.132360] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b841514-748f-497d-80cf-73f04fbd9455 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.146252] env[61991]: DEBUG nova.compute.provider_tree [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 502.654230] env[61991]: DEBUG nova.scheduler.client.report [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 502.665698] env[61991]: ERROR nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 502.665698] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 502.665698] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 502.665698] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 502.665698] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 502.665698] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 502.665698] env[61991]: ERROR nova.compute.manager raise self.value [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 502.665698] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 502.665698] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 502.665698] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 502.667453] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 502.667453] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 502.667453] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 502.667453] env[61991]: ERROR nova.compute.manager [ 502.667453] env[61991]: Traceback (most recent call last): [ 502.667453] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 502.667453] env[61991]: listener.cb(fileno) [ 502.667453] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.667453] env[61991]: result = function(*args, **kwargs) [ 502.667453] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 502.667453] env[61991]: return func(*args, **kwargs) [ 502.667453] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 502.667453] env[61991]: raise e [ 502.667453] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 502.667453] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 502.667453] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 502.667453] env[61991]: created_port_ids = self._update_ports_for_instance( [ 502.667453] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 502.667453] env[61991]: with excutils.save_and_reraise_exception(): [ 502.667453] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 502.667453] env[61991]: self.force_reraise() [ 502.667453] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 502.667453] env[61991]: raise self.value [ 502.667453] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 502.667453] env[61991]: updated_port = self._update_port( [ 502.667453] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 502.667453] env[61991]: _ensure_no_port_binding_failure(port) [ 502.667453] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 502.667453] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 502.668285] env[61991]: nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 502.668285] env[61991]: Removing descriptor: 15 [ 502.668285] env[61991]: ERROR nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Traceback (most recent call last): [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] yield resources [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self.driver.spawn(context, instance, image_meta, [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 502.668285] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] vm_ref = self.build_virtual_machine(instance, [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] for vif in network_info: [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return self._sync_wrapper(fn, *args, **kwargs) [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self.wait() [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self[:] = self._gt.wait() [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return self._exit_event.wait() [ 502.668629] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] result = hub.switch() [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return self.greenlet.switch() [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] result = function(*args, **kwargs) [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return func(*args, **kwargs) [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] raise e [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] nwinfo = self.network_api.allocate_for_instance( [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 502.668966] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] created_port_ids = self._update_ports_for_instance( [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] with excutils.save_and_reraise_exception(): [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self.force_reraise() [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] raise self.value [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] updated_port = self._update_port( [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] _ensure_no_port_binding_failure(port) [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 502.669319] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] raise exception.PortBindingFailed(port_id=port['id']) [ 502.669709] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 502.669709] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] [ 502.669709] env[61991]: INFO nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Terminating instance [ 502.672366] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.672671] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.672758] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 502.942279] env[61991]: DEBUG nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 502.977063] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 502.977063] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 502.977063] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 502.977240] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 502.977240] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 502.977240] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 502.977240] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 502.977240] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 502.977377] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 502.977377] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 502.977377] env[61991]: DEBUG nova.virt.hardware [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 502.978349] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db66b982-7561-4fc6-9cb2-7a5ed72b558d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.988078] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2387e9a-1096-494c-9e89-66b34884d8fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.003453] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 503.019873] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 503.020228] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4bb802b-9054-43e3-a5fd-bd2e28683c63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.032419] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Created folder: OpenStack in parent group-v4. [ 503.032465] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Creating folder: Project (822784945e2744df8140cdf5d0733728). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 503.032674] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad076444-9e7f-45b3-b000-64fb06614c17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.043390] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Created folder: Project (822784945e2744df8140cdf5d0733728) in parent group-v246753. [ 503.043542] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Creating folder: Instances. Parent ref: group-v246754. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 503.043768] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f922b27-911a-4018-a711-e1ee2e74db90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.054889] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Created folder: Instances in parent group-v246754. [ 503.055171] env[61991]: DEBUG oslo.service.loopingcall [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 503.055354] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 503.055542] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65042fca-3e64-4251-8fc1-8051ae62f526 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.073481] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 503.073481] env[61991]: value = "task-1129254" [ 503.073481] env[61991]: _type = "Task" [ 503.073481] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 503.081896] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129254, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 503.160184] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 503.160184] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 503.164512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.851s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.166646] env[61991]: INFO nova.compute.claims [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 503.266853] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 503.379921] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquiring lock "8014b7df-35e5-403c-b0dc-b49602d59060" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.380242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Lock "8014b7df-35e5-403c-b0dc-b49602d59060" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.562458] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.585565] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129254, 'name': CreateVM_Task, 'duration_secs': 0.357449} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 503.585794] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 503.587012] env[61991]: DEBUG oslo_vmware.service [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f356af-6f50-40b9-871d-4bd33653ccc7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.594191] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 503.594364] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 503.595563] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 503.596512] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d09b8c41-7053-4589-b9d9-dcec95f0a916 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.604046] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 503.604046] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527bd5a2-c23b-3093-bf3e-d5859ee8dbe3" [ 503.604046] env[61991]: _type = "Task" [ 503.604046] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 503.613060] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527bd5a2-c23b-3093-bf3e-d5859ee8dbe3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 503.671874] env[61991]: DEBUG nova.compute.utils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 503.679018] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 503.679018] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 503.718325] env[61991]: DEBUG nova.policy [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e6810d3aced4a44bbfe9ba49270bc35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a18b2a73f7643e1b89c9660a770e9da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 503.885075] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 503.901492] env[61991]: DEBUG nova.compute.manager [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Received event network-changed-d093b2e7-784c-4581-bc31-9c6be567a256 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 503.902719] env[61991]: DEBUG nova.compute.manager [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Refreshing instance network info cache due to event network-changed-d093b2e7-784c-4581-bc31-9c6be567a256. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 503.903517] env[61991]: DEBUG oslo_concurrency.lockutils [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] Acquiring lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 504.068490] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 504.068490] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 504.068490] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 504.068490] env[61991]: DEBUG oslo_concurrency.lockutils [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] Acquired lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 504.068490] env[61991]: DEBUG nova.network.neutron [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Refreshing network info cache for port d093b2e7-784c-4581-bc31-9c6be567a256 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 504.069197] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24819985-d2f7-4bc1-840a-4c5e0b8bb92c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.082251] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee26f71-d661-4e34-a7bd-1708267ec595 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.106625] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea could not be found. [ 504.106893] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 504.107338] env[61991]: INFO nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 504.107572] env[61991]: DEBUG oslo.service.loopingcall [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 504.112126] env[61991]: DEBUG nova.compute.manager [-] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 504.112250] env[61991]: DEBUG nova.network.neutron [-] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 504.123502] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 504.127023] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 504.127023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 504.127023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 504.127023] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 504.127023] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb7b74cd-6e0b-430d-b58f-465d648f7356 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.133170] env[61991]: DEBUG nova.network.neutron [-] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 504.143584] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 504.143757] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 504.144652] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bf8ebc-c858-4438-8539-39fd1a5fb309 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.152718] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Successfully created port: 720150e6-5de7-4610-af0c-8b954d9bd1b1 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 504.154621] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ec15786-5c22-4bb2-a939-909203afa2f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.165395] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 504.165395] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5254f675-b939-b203-852b-842524598e09" [ 504.165395] env[61991]: _type = "Task" [ 504.165395] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 504.174280] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Preparing fetch location {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 504.174525] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Creating directory with path [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 504.174808] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6968dcc-7de1-46f6-884d-da2ee11fc303 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.182729] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 504.205383] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Created directory with path [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 504.205485] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Fetch image to [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 504.205860] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Downloading image file data 254d700f-2f5a-49a3-8762-cec07162124a to [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk on the data store datastore2 {{(pid=61991) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 504.206500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375ef2f4-6d75-43b4-9c83-be8ba1f749b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.214901] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5161de17-c3c3-4ab2-9ccd-31ee73f03594 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.228249] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48ad66f-9cc6-42f6-921b-ce0ccc87dc7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.264677] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef9342d-1db2-44e5-a4cc-62a0eac1b71d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.270235] env[61991]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4e63329f-6d11-48f3-9e19-07ba11367312 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.353584] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28af4025-07eb-410c-a33d-a69da4a28f2a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.366825] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Downloading image file data 254d700f-2f5a-49a3-8762-cec07162124a to the data store datastore2 {{(pid=61991) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 504.373674] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4047783-84c7-4d58-9d38-d7589d0e506e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.418543] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fccbe5-7b9b-42ba-a6d9-162cee9e5931 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.433197] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ec8238-8172-4c7a-96fd-bd11b69aabb2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.439418] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.451574] env[61991]: DEBUG nova.compute.provider_tree [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 504.491740] env[61991]: DEBUG oslo_vmware.rw_handles [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61991) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 504.620615] env[61991]: DEBUG nova.network.neutron [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 504.639045] env[61991]: DEBUG nova.network.neutron [-] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 504.792355] env[61991]: DEBUG nova.network.neutron [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 504.956037] env[61991]: DEBUG nova.scheduler.client.report [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 505.117105] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquiring lock "a12782a9-3730-4c2f-826e-d1e68d2c182b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.118397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Lock "a12782a9-3730-4c2f-826e-d1e68d2c182b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.144662] env[61991]: INFO nova.compute.manager [-] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Took 1.03 seconds to deallocate network for instance. [ 505.149127] env[61991]: DEBUG nova.compute.claims [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 505.149127] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.157705] env[61991]: DEBUG oslo_vmware.rw_handles [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Completed reading data from the image iterator. {{(pid=61991) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 505.158375] env[61991]: DEBUG oslo_vmware.rw_handles [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 505.197027] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 505.219277] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 505.220012] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 505.220012] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 505.220012] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 505.220167] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 505.220403] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 505.220720] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 505.221327] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 505.221734] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 505.222892] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 505.222892] env[61991]: DEBUG nova.virt.hardware [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 505.223358] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d6ff90-bfae-46b9-9a2e-707317092829 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.232889] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55d81d9-af0d-403f-a213-2e0d49593b17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.295983] env[61991]: DEBUG oslo_concurrency.lockutils [req-c3082b42-4490-4a7d-a983-7d76a3deb667 req-9820294b-0ce9-4218-ae1e-f70fe5ec3452 service nova] Releasing lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 505.302992] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Downloaded image file data 254d700f-2f5a-49a3-8762-cec07162124a to vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk on the data store datastore2 {{(pid=61991) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 505.305442] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Caching image {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 505.305886] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Copying Virtual Disk [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk to [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 505.306186] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c1149fc-57d5-4c50-b24b-1e6e13d42fc3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.315041] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 505.315041] env[61991]: value = "task-1129255" [ 505.315041] env[61991]: _type = "Task" [ 505.315041] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 505.323779] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129255, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.332401] env[61991]: ERROR nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 505.332401] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 505.332401] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 505.332401] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 505.332401] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.332401] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.332401] env[61991]: ERROR nova.compute.manager raise self.value [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 505.332401] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 505.332401] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.332401] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 505.332850] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.332850] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 505.332850] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 505.332850] env[61991]: ERROR nova.compute.manager [ 505.332850] env[61991]: Traceback (most recent call last): [ 505.332850] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 505.332850] env[61991]: listener.cb(fileno) [ 505.332850] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 505.332850] env[61991]: result = function(*args, **kwargs) [ 505.332850] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 505.332850] env[61991]: return func(*args, **kwargs) [ 505.332850] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 505.332850] env[61991]: raise e [ 505.332850] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 505.332850] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 505.332850] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 505.332850] env[61991]: created_port_ids = self._update_ports_for_instance( [ 505.332850] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 505.332850] env[61991]: with excutils.save_and_reraise_exception(): [ 505.332850] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.332850] env[61991]: self.force_reraise() [ 505.332850] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.332850] env[61991]: raise self.value [ 505.332850] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 505.332850] env[61991]: updated_port = self._update_port( [ 505.332850] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.332850] env[61991]: _ensure_no_port_binding_failure(port) [ 505.332850] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.332850] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 505.333626] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 505.333626] env[61991]: Removing descriptor: 15 [ 505.333626] env[61991]: ERROR nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Traceback (most recent call last): [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] yield resources [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self.driver.spawn(context, instance, image_meta, [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 505.333626] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] vm_ref = self.build_virtual_machine(instance, [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] vif_infos = vmwarevif.get_vif_info(self._session, [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] for vif in network_info: [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return self._sync_wrapper(fn, *args, **kwargs) [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self.wait() [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self[:] = self._gt.wait() [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return self._exit_event.wait() [ 505.333940] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] result = hub.switch() [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return self.greenlet.switch() [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] result = function(*args, **kwargs) [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return func(*args, **kwargs) [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] raise e [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] nwinfo = self.network_api.allocate_for_instance( [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 505.334293] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] created_port_ids = self._update_ports_for_instance( [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] with excutils.save_and_reraise_exception(): [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self.force_reraise() [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] raise self.value [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] updated_port = self._update_port( [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] _ensure_no_port_binding_failure(port) [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.334642] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] raise exception.PortBindingFailed(port_id=port['id']) [ 505.334980] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 505.334980] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] [ 505.334980] env[61991]: INFO nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Terminating instance [ 505.336969] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 505.337172] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 505.337353] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 505.461844] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.462445] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 505.466367] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.031s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.467839] env[61991]: INFO nova.compute.claims [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 505.623779] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 505.828223] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129255, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.869153] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 505.976109] env[61991]: DEBUG nova.compute.utils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 505.979927] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 505.981335] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 506.089264] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 506.092435] env[61991]: DEBUG nova.policy [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ea75fdec6c24525a19f8ec7867010b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'deec43f5fee94ec9a301c52f57c90b19', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 506.154722] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.281944] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquiring lock "5a9d879c-3220-4a16-b119-2ce6949f273f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.282214] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Lock "5a9d879c-3220-4a16-b119-2ce6949f273f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.326440] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129255, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.664756} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 506.326926] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Copied Virtual Disk [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk to [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 506.327282] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleting the datastore file [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 506.328614] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1d140e1-6a12-4a68-91e1-6dd1e76957eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.334426] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 506.334426] env[61991]: value = "task-1129256" [ 506.334426] env[61991]: _type = "Task" [ 506.334426] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.342772] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129256, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.415770] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquiring lock "c8564e62-162a-4df1-a9ac-8826879cec38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.416035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Lock "c8564e62-162a-4df1-a9ac-8826879cec38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.481428] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 506.595106] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 506.596661] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 506.596661] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 506.596661] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-faad3367-24ff-4a2f-b035-3afb96ca558d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.616775] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d102a3-1e5e-41c7-9432-c79e40737d99 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.644249] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f32fee48-de95-4ade-9ed0-092786d3f39c could not be found. [ 506.644718] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 506.645063] env[61991]: INFO nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 506.645557] env[61991]: DEBUG oslo.service.loopingcall [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 506.652119] env[61991]: DEBUG nova.compute.manager [-] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 506.652363] env[61991]: DEBUG nova.network.neutron [-] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 506.688640] env[61991]: DEBUG nova.network.neutron [-] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 506.710134] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb474242-cb41-4d38-95fa-3fce5b339f25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.718378] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e995f21-c4c0-4712-9ff3-fc3b8606a568 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.751363] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b78852-97ba-41f0-b794-9fade9ef9fec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.758976] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3447c5a6-c6cd-4f1d-8a22-3b45e74fdae4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.774823] env[61991]: DEBUG nova.compute.provider_tree [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 506.777569] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Successfully created port: 3edcffb6-8ae2-4953-a164-03156abbf10c {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 506.787926] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 506.845726] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129256, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022452} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 506.846113] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 506.846508] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Moving file from [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a/254d700f-2f5a-49a3-8762-cec07162124a to [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a. {{(pid=61991) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 506.846866] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-0bd5108a-a4ef-4f05-825f-a20937eb3068 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.854958] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 506.854958] env[61991]: value = "task-1129257" [ 506.854958] env[61991]: _type = "Task" [ 506.854958] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.863714] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129257, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.918533] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 507.138138] env[61991]: DEBUG nova.compute.manager [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Received event network-vif-deleted-d093b2e7-784c-4581-bc31-9c6be567a256 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 507.138138] env[61991]: DEBUG nova.compute.manager [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Received event network-changed-720150e6-5de7-4610-af0c-8b954d9bd1b1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 507.138138] env[61991]: DEBUG nova.compute.manager [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Refreshing instance network info cache due to event network-changed-720150e6-5de7-4610-af0c-8b954d9bd1b1. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 507.138138] env[61991]: DEBUG oslo_concurrency.lockutils [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] Acquiring lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 507.138549] env[61991]: DEBUG oslo_concurrency.lockutils [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] Acquired lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 507.138549] env[61991]: DEBUG nova.network.neutron [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Refreshing network info cache for port 720150e6-5de7-4610-af0c-8b954d9bd1b1 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 507.191376] env[61991]: DEBUG nova.network.neutron [-] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 507.281163] env[61991]: DEBUG nova.scheduler.client.report [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 507.314340] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.368837] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129257, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.022188} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.369169] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] File moved {{(pid=61991) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 507.369367] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Cleaning up location [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 507.369525] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleting the datastore file [datastore2] vmware_temp/c7b8cc68-1e40-4ac1-8b17-4f8cc46a512a {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 507.369773] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b03ba1ab-c6b8-4dbf-ab6b-6d2f0efc57ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.377643] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 507.377643] env[61991]: value = "task-1129258" [ 507.377643] env[61991]: _type = "Task" [ 507.377643] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.385532] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.451934] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.491766] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 507.531218] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 507.531449] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 507.531602] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 507.532658] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 507.532825] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 507.533492] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 507.533492] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 507.533492] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 507.535288] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 507.535548] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 507.535645] env[61991]: DEBUG nova.virt.hardware [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 507.537032] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1198347-101e-44c3-a9d2-35e215375df3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.549224] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2dfc2e-d3ab-4d5e-941c-5efab2c1aabc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.694202] env[61991]: INFO nova.compute.manager [-] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Took 1.04 seconds to deallocate network for instance. [ 507.699064] env[61991]: DEBUG nova.compute.claims [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 507.699064] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.723916] env[61991]: DEBUG nova.network.neutron [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 507.789027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 507.789027] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 507.790627] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.964s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.793375] env[61991]: INFO nova.compute.claims [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 507.889028] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024843} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.889519] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 507.890442] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbd5be2f-1961-4334-9d2e-7e15167f119e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.896189] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 507.896189] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5277f1bf-5166-7546-1d7c-800c0b851348" [ 507.896189] env[61991]: _type = "Task" [ 507.896189] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.905445] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5277f1bf-5166-7546-1d7c-800c0b851348, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.979598] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 507.979940] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 507.980151] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 507.980259] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 508.151719] env[61991]: DEBUG nova.network.neutron [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 508.300244] env[61991]: DEBUG nova.compute.utils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 508.307058] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 508.307058] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 508.408341] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5277f1bf-5166-7546-1d7c-800c0b851348, 'name': SearchDatastore_Task, 'duration_secs': 0.009123} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 508.408704] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 508.408986] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] d978d68a-0709-4a32-bbaf-3bfa006be85c/d978d68a-0709-4a32-bbaf-3bfa006be85c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 508.409547] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-202960ce-e4e9-4f45-925e-8895f2272a5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.416818] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 508.416818] env[61991]: value = "task-1129259" [ 508.416818] env[61991]: _type = "Task" [ 508.416818] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.425372] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.486997] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 508.487270] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 508.488630] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 508.488777] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 508.488988] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 508.489056] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 508.489160] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Didn't find any instances for network info cache update. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 508.489384] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.489640] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.489828] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.490055] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.490411] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.490659] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.490811] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 508.490959] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.531034] env[61991]: DEBUG nova.policy [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3def0d3763441d8a0f1b248f9b29c14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '12da0c1e6ea741a5b4e84c58a9b09ce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 508.655301] env[61991]: DEBUG oslo_concurrency.lockutils [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] Releasing lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 508.655543] env[61991]: DEBUG nova.compute.manager [req-c4721922-b795-450e-a6d9-3c3363a01aa7 req-8df570be-c5a5-44f0-be6d-4d5631ac10e3 service nova] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Received event network-vif-deleted-720150e6-5de7-4610-af0c-8b954d9bd1b1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 508.811515] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 508.932590] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129259, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.994350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.004700] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90db4239-ee77-4224-95d6-e95cdada5e91 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.012874] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5cbcbe-90bd-462c-abbe-d0321e5b94d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.045298] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a9eae2-5efc-4f39-89fb-d0cee538f3fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.052873] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328e0bb3-2253-49ff-ad2b-3fc130e79cbb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.067574] env[61991]: DEBUG nova.compute.provider_tree [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 509.432157] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592733} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.432157] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] d978d68a-0709-4a32-bbaf-3bfa006be85c/d978d68a-0709-4a32-bbaf-3bfa006be85c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 509.434441] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 509.437596] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fbf1501-ac7c-437a-8df3-29fd7a79d95e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.444315] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 509.444315] env[61991]: value = "task-1129260" [ 509.444315] env[61991]: _type = "Task" [ 509.444315] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.454978] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.501127] env[61991]: ERROR nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 509.501127] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 509.501127] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 509.501127] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 509.501127] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 509.501127] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 509.501127] env[61991]: ERROR nova.compute.manager raise self.value [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 509.501127] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 509.501127] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 509.501127] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 509.501534] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 509.501534] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 509.501534] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 509.501534] env[61991]: ERROR nova.compute.manager [ 509.501534] env[61991]: Traceback (most recent call last): [ 509.501534] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 509.501534] env[61991]: listener.cb(fileno) [ 509.501534] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 509.501534] env[61991]: result = function(*args, **kwargs) [ 509.501534] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 509.501534] env[61991]: return func(*args, **kwargs) [ 509.501534] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 509.501534] env[61991]: raise e [ 509.501534] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 509.501534] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 509.501534] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 509.501534] env[61991]: created_port_ids = self._update_ports_for_instance( [ 509.501534] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 509.501534] env[61991]: with excutils.save_and_reraise_exception(): [ 509.501534] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 509.501534] env[61991]: self.force_reraise() [ 509.501534] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 509.501534] env[61991]: raise self.value [ 509.501534] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 509.501534] env[61991]: updated_port = self._update_port( [ 509.501534] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 509.501534] env[61991]: _ensure_no_port_binding_failure(port) [ 509.501534] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 509.501534] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 509.502305] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 509.502305] env[61991]: Removing descriptor: 17 [ 509.502305] env[61991]: ERROR nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Traceback (most recent call last): [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] yield resources [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self.driver.spawn(context, instance, image_meta, [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self._vmops.spawn(context, instance, image_meta, injected_files, [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 509.502305] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] vm_ref = self.build_virtual_machine(instance, [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] vif_infos = vmwarevif.get_vif_info(self._session, [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] for vif in network_info: [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return self._sync_wrapper(fn, *args, **kwargs) [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self.wait() [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self[:] = self._gt.wait() [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return self._exit_event.wait() [ 509.502617] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] result = hub.switch() [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return self.greenlet.switch() [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] result = function(*args, **kwargs) [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return func(*args, **kwargs) [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] raise e [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] nwinfo = self.network_api.allocate_for_instance( [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 509.503560] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] created_port_ids = self._update_ports_for_instance( [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] with excutils.save_and_reraise_exception(): [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self.force_reraise() [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] raise self.value [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] updated_port = self._update_port( [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] _ensure_no_port_binding_failure(port) [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 509.504232] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] raise exception.PortBindingFailed(port_id=port['id']) [ 509.504653] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 509.504653] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] [ 509.504653] env[61991]: INFO nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Terminating instance [ 509.506010] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquiring lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 509.506213] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquired lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 509.506442] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 509.573637] env[61991]: DEBUG nova.scheduler.client.report [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 509.826803] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 509.865019] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 509.865019] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 509.865019] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 509.865363] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 509.865363] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 509.865363] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 509.865363] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 509.865363] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 509.865591] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 509.865591] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 509.865591] env[61991]: DEBUG nova.virt.hardware [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 509.865888] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12238a7c-7e5b-4ce1-9064-5bd31633e099 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.874694] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c45aea1-f006-4bd9-b64d-b975824eb660 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.953643] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066109} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.954118] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 509.955008] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87a3371-d442-4754-ac6c-8e14fac9b2fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.980600] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] d978d68a-0709-4a32-bbaf-3bfa006be85c/d978d68a-0709-4a32-bbaf-3bfa006be85c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 509.980924] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74802b84-e83c-4b08-8ca7-9666783b183f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.001081] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 510.001081] env[61991]: value = "task-1129261" [ 510.001081] env[61991]: _type = "Task" [ 510.001081] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.009127] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129261, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.041989] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Successfully created port: d8d7a50c-23e7-422e-9171-8acd4d4e9433 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 510.077187] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.077776] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 510.080529] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.641s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.082337] env[61991]: INFO nova.compute.claims [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.118962] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 510.287639] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 510.312217] env[61991]: DEBUG nova.compute.manager [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Received event network-changed-3edcffb6-8ae2-4953-a164-03156abbf10c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 510.312217] env[61991]: DEBUG nova.compute.manager [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Refreshing instance network info cache due to event network-changed-3edcffb6-8ae2-4953-a164-03156abbf10c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 510.312217] env[61991]: DEBUG oslo_concurrency.lockutils [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] Acquiring lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.516161] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129261, 'name': ReconfigVM_Task, 'duration_secs': 0.352683} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.517421] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Reconfigured VM instance instance-00000002 to attach disk [datastore2] d978d68a-0709-4a32-bbaf-3bfa006be85c/d978d68a-0709-4a32-bbaf-3bfa006be85c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 510.519264] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-edab2611-46dd-4d5e-9188-d77416ebbffc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.526728] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 510.526728] env[61991]: value = "task-1129262" [ 510.526728] env[61991]: _type = "Task" [ 510.526728] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.538560] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129262, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.587252] env[61991]: DEBUG nova.compute.utils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 510.592617] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 510.592617] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 510.778040] env[61991]: DEBUG nova.policy [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17eb886fb2df46a085b1e7d8812c4890', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd184a2c7d9f84e0c8464eae053e0539c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 510.791486] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Releasing lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.791837] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 510.792037] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 510.792434] env[61991]: DEBUG oslo_concurrency.lockutils [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] Acquired lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 510.792587] env[61991]: DEBUG nova.network.neutron [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Refreshing network info cache for port 3edcffb6-8ae2-4953-a164-03156abbf10c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 510.797559] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8f71ef4-5e1f-4338-943a-369e279c3dcd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.810883] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad76bb09-4622-4663-97b2-c8a7655fbf95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.837090] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607 could not be found. [ 510.838136] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 510.838136] env[61991]: INFO nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Took 0.05 seconds to destroy the instance on the hypervisor. [ 510.838136] env[61991]: DEBUG oslo.service.loopingcall [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 510.838136] env[61991]: DEBUG nova.compute.manager [-] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 510.838136] env[61991]: DEBUG nova.network.neutron [-] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 510.864815] env[61991]: DEBUG nova.network.neutron [-] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 511.037490] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129262, 'name': Rename_Task, 'duration_secs': 0.138088} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.037756] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 511.038095] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73ea90e6-9f3c-4a6c-9a01-a5e45365ef07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.044907] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 511.044907] env[61991]: value = "task-1129263" [ 511.044907] env[61991]: _type = "Task" [ 511.044907] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.054273] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.096244] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 511.283847] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquiring lock "ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.284317] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Lock "ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.367829] env[61991]: DEBUG nova.network.neutron [-] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.381146] env[61991]: DEBUG nova.network.neutron [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 511.435462] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d298bba8-60f4-4ac7-960e-46bcd5faf7b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.444374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9a0c02-5810-422c-b980-20c43fa3b1a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.477585] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefe9d63-8bcb-454a-be63-4c0b3e1363f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.486848] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d10903-0221-4456-8584-a5f38ef14418 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.502140] env[61991]: DEBUG nova.compute.provider_tree [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 511.555940] env[61991]: DEBUG oslo_vmware.api [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129263, 'name': PowerOnVM_Task, 'duration_secs': 0.476853} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.559199] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 511.559199] env[61991]: INFO nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Took 8.61 seconds to spawn the instance on the hypervisor. [ 511.559199] env[61991]: DEBUG nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 511.559199] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e417b86c-fb1e-47d4-9f8c-99d5b8d09497 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.639977] env[61991]: DEBUG nova.network.neutron [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.878784] env[61991]: INFO nova.compute.manager [-] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Took 1.04 seconds to deallocate network for instance. [ 511.880325] env[61991]: DEBUG nova.compute.claims [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 511.880654] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.006406] env[61991]: DEBUG nova.scheduler.client.report [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 512.078623] env[61991]: INFO nova.compute.manager [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Took 14.58 seconds to build instance. [ 512.115285] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 512.145444] env[61991]: DEBUG oslo_concurrency.lockutils [req-54532f22-cd5a-470e-84ff-7ba8d6f905e0 req-048930fc-14e4-402c-b380-cf3ec3d20807 service nova] Releasing lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.145444] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 512.146809] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 512.147025] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 512.147250] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 512.147398] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 512.147538] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 512.147737] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 512.147916] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 512.148287] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 512.148363] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 512.148578] env[61991]: DEBUG nova.virt.hardware [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 512.149842] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eeb3e03-5720-4f32-87af-3fc7a66bad4d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.158475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfe6b06-c0cf-44ba-af7c-0c36eac9d5ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.195683] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Successfully created port: d92d491b-f4d1-494c-aa31-804bf04a46bc {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 512.511730] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.512525] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 512.516187] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.368s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.587525] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920c5c47-fcff-4cb1-8e56-b5794b3486d3 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "d978d68a-0709-4a32-bbaf-3bfa006be85c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.106s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.025266] env[61991]: DEBUG nova.compute.utils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.032474] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 513.032474] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 513.092739] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 513.219217] env[61991]: DEBUG nova.policy [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c381ac0afb094bd099295bb89b13b5c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f407b440184f4b39bf115737d0e25f5b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 513.228139] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquiring lock "3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.228365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Lock "3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.251108] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96823c23-6d7b-448d-8742-93bc9e75696b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.259583] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d68c36-3a38-4da4-803d-a68c1ce34da5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.303588] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a24ac0d-6a47-4a67-873e-69c5627a5295 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.313233] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c151691e-5d5c-4916-a2ce-a9374a479b41 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.327653] env[61991]: DEBUG nova.compute.provider_tree [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 513.532554] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 513.624643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.834133] env[61991]: DEBUG nova.scheduler.client.report [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 514.035376] env[61991]: DEBUG nova.compute.manager [req-8942977f-4f44-49f9-bc45-ae978d7a47eb req-fa42c312-c96a-49aa-9981-18f1758a6916 service nova] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Received event network-vif-deleted-3edcffb6-8ae2-4953-a164-03156abbf10c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 514.347027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.829s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.347027] env[61991]: ERROR nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Traceback (most recent call last): [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self.driver.spawn(context, instance, image_meta, [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 514.347027] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] vm_ref = self.build_virtual_machine(instance, [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] for vif in network_info: [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return self._sync_wrapper(fn, *args, **kwargs) [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self.wait() [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self[:] = self._gt.wait() [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return self._exit_event.wait() [ 514.347633] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] result = hub.switch() [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return self.greenlet.switch() [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] result = function(*args, **kwargs) [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] return func(*args, **kwargs) [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] raise e [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] nwinfo = self.network_api.allocate_for_instance( [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 514.348184] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] created_port_ids = self._update_ports_for_instance( [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] with excutils.save_and_reraise_exception(): [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] self.force_reraise() [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] raise self.value [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] updated_port = self._update_port( [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] _ensure_no_port_binding_failure(port) [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 514.348876] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] raise exception.PortBindingFailed(port_id=port['id']) [ 514.349402] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] nova.exception.PortBindingFailed: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. [ 514.349402] env[61991]: ERROR nova.compute.manager [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] [ 514.354411] env[61991]: DEBUG nova.compute.utils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 514.354411] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.198s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.356974] env[61991]: INFO nova.compute.claims [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.366368] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Build of instance 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea was re-scheduled: Binding failed for port d093b2e7-784c-4581-bc31-9c6be567a256, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 514.367339] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 514.367836] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.368221] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.369103] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 514.462150] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquiring lock "38d3d1bf-8607-494a-a991-0bc4de3d7522" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.462502] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Lock "38d3d1bf-8607-494a-a991-0bc4de3d7522" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.545192] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 514.576683] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 514.577559] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 514.577903] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 514.578244] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 514.578591] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 514.578901] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 514.579225] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 514.579466] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 514.579690] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 514.579906] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 514.580165] env[61991]: DEBUG nova.virt.hardware [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 514.581095] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888ac327-8655-4349-955d-77d75c3b974d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.589747] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f02839-c6f6-4cfe-854a-9a85c87e5040 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.776600] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Successfully created port: 1b61fba4-f7af-44ad-9dbc-196fd7794dc2 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 514.944383] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 515.022082] env[61991]: ERROR nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 515.022082] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.022082] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 515.022082] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 515.022082] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.022082] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.022082] env[61991]: ERROR nova.compute.manager raise self.value [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 515.022082] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 515.022082] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.022082] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 515.022550] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.022550] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 515.022550] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 515.022550] env[61991]: ERROR nova.compute.manager [ 515.022550] env[61991]: Traceback (most recent call last): [ 515.022550] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 515.022550] env[61991]: listener.cb(fileno) [ 515.022550] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.022550] env[61991]: result = function(*args, **kwargs) [ 515.022550] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.022550] env[61991]: return func(*args, **kwargs) [ 515.022550] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 515.022550] env[61991]: raise e [ 515.022550] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.022550] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 515.022550] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 515.022550] env[61991]: created_port_ids = self._update_ports_for_instance( [ 515.022550] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 515.022550] env[61991]: with excutils.save_and_reraise_exception(): [ 515.022550] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.022550] env[61991]: self.force_reraise() [ 515.022550] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.022550] env[61991]: raise self.value [ 515.022550] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 515.022550] env[61991]: updated_port = self._update_port( [ 515.022550] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.022550] env[61991]: _ensure_no_port_binding_failure(port) [ 515.022550] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.022550] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 515.023730] env[61991]: nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 515.023730] env[61991]: Removing descriptor: 15 [ 515.023730] env[61991]: ERROR nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Traceback (most recent call last): [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] yield resources [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self.driver.spawn(context, instance, image_meta, [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 515.023730] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] vm_ref = self.build_virtual_machine(instance, [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] vif_infos = vmwarevif.get_vif_info(self._session, [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] for vif in network_info: [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return self._sync_wrapper(fn, *args, **kwargs) [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self.wait() [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self[:] = self._gt.wait() [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return self._exit_event.wait() [ 515.024069] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] result = hub.switch() [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return self.greenlet.switch() [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] result = function(*args, **kwargs) [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return func(*args, **kwargs) [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] raise e [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] nwinfo = self.network_api.allocate_for_instance( [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 515.024415] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] created_port_ids = self._update_ports_for_instance( [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] with excutils.save_and_reraise_exception(): [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self.force_reraise() [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] raise self.value [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] updated_port = self._update_port( [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] _ensure_no_port_binding_failure(port) [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.024793] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] raise exception.PortBindingFailed(port_id=port['id']) [ 515.025121] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 515.025121] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] [ 515.025121] env[61991]: INFO nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Terminating instance [ 515.026234] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.026465] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquired lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.026687] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 515.261765] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.576629] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 515.610483] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24f1516-39eb-406e-a753-955d8278e731 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.622026] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487eabce-e024-4dc4-82d2-e14796619c9b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.655124] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b87ccb-e9aa-4901-b137-c763bff95be0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.663185] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7519898d-edb6-4d23-b0f3-d7d42fc762eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.680894] env[61991]: DEBUG nova.compute.provider_tree [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.689897] env[61991]: DEBUG nova.compute.manager [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Received event network-changed-d8d7a50c-23e7-422e-9171-8acd4d4e9433 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 515.690121] env[61991]: DEBUG nova.compute.manager [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Refreshing instance network info cache due to event network-changed-d8d7a50c-23e7-422e-9171-8acd4d4e9433. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 515.691111] env[61991]: DEBUG oslo_concurrency.lockutils [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] Acquiring lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.764821] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.768020] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 515.768020] env[61991]: DEBUG nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 515.768020] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 515.879617] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 515.889035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquiring lock "8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.889035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Lock "8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.950073] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.187085] env[61991]: DEBUG nova.scheduler.client.report [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.382697] env[61991]: DEBUG nova.network.neutron [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.389157] env[61991]: DEBUG nova.compute.manager [None req-fb90e451-b577-42ec-b789-8f89c80aa63a tempest-ServerDiagnosticsV248Test-85672924 tempest-ServerDiagnosticsV248Test-85672924-project-admin] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 516.389157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8c7d19-18ba-4f53-a97e-032760c21429 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.397430] env[61991]: INFO nova.compute.manager [None req-fb90e451-b577-42ec-b789-8f89c80aa63a tempest-ServerDiagnosticsV248Test-85672924 tempest-ServerDiagnosticsV248Test-85672924-project-admin] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Retrieving diagnostics [ 516.399512] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40d12a4-71ff-407b-a0c5-a93cde217d72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.453446] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Releasing lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.453722] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 516.454193] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 516.454346] env[61991]: DEBUG oslo_concurrency.lockutils [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] Acquired lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.455432] env[61991]: DEBUG nova.network.neutron [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Refreshing network info cache for port d8d7a50c-23e7-422e-9171-8acd4d4e9433 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 516.456121] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04b8a25e-d3d8-485a-ac59-2ba4f0af5bc3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.466582] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8c3e05-f33f-416f-810b-8f546eff7177 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.495334] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e4e28fbe-9f51-4404-9d0c-c0d80a11cb38 could not be found. [ 516.495658] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 516.495930] env[61991]: INFO nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Took 0.04 seconds to destroy the instance on the hypervisor. [ 516.496241] env[61991]: DEBUG oslo.service.loopingcall [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 516.496554] env[61991]: DEBUG nova.compute.manager [-] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 516.496708] env[61991]: DEBUG nova.network.neutron [-] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 516.566105] env[61991]: DEBUG nova.network.neutron [-] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 516.568841] env[61991]: ERROR nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 516.568841] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 516.568841] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 516.568841] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 516.568841] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 516.568841] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 516.568841] env[61991]: ERROR nova.compute.manager raise self.value [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 516.568841] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 516.568841] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 516.568841] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 516.569308] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 516.569308] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 516.569308] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 516.569308] env[61991]: ERROR nova.compute.manager [ 516.569418] env[61991]: Traceback (most recent call last): [ 516.569470] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 516.569470] env[61991]: listener.cb(fileno) [ 516.569470] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 516.569470] env[61991]: result = function(*args, **kwargs) [ 516.569470] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 516.569470] env[61991]: return func(*args, **kwargs) [ 516.569470] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 516.569470] env[61991]: raise e [ 516.569694] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 516.569694] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 516.569694] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 516.569694] env[61991]: created_port_ids = self._update_ports_for_instance( [ 516.569694] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 516.569694] env[61991]: with excutils.save_and_reraise_exception(): [ 516.569694] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 516.569694] env[61991]: self.force_reraise() [ 516.569694] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 516.569694] env[61991]: raise self.value [ 516.569694] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 516.569694] env[61991]: updated_port = self._update_port( [ 516.569694] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 516.569694] env[61991]: _ensure_no_port_binding_failure(port) [ 516.569694] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 516.569694] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 516.569694] env[61991]: nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 516.569694] env[61991]: Removing descriptor: 17 [ 516.570993] env[61991]: ERROR nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Traceback (most recent call last): [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] yield resources [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self.driver.spawn(context, instance, image_meta, [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] vm_ref = self.build_virtual_machine(instance, [ 516.570993] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] vif_infos = vmwarevif.get_vif_info(self._session, [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] for vif in network_info: [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return self._sync_wrapper(fn, *args, **kwargs) [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self.wait() [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self[:] = self._gt.wait() [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return self._exit_event.wait() [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 516.571351] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] result = hub.switch() [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return self.greenlet.switch() [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] result = function(*args, **kwargs) [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return func(*args, **kwargs) [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] raise e [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] nwinfo = self.network_api.allocate_for_instance( [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] created_port_ids = self._update_ports_for_instance( [ 516.571675] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] with excutils.save_and_reraise_exception(): [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self.force_reraise() [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] raise self.value [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] updated_port = self._update_port( [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] _ensure_no_port_binding_failure(port) [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] raise exception.PortBindingFailed(port_id=port['id']) [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 516.571988] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] [ 516.572383] env[61991]: INFO nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Terminating instance [ 516.573863] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquiring lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.574174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquired lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.574404] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 516.695027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.695027] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 516.700389] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.385s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.700725] env[61991]: INFO nova.compute.claims [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 516.890438] env[61991]: INFO nova.compute.manager [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea] Took 1.12 seconds to deallocate network for instance. [ 516.991020] env[61991]: DEBUG nova.network.neutron [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 517.073469] env[61991]: DEBUG nova.network.neutron [-] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.126549] env[61991]: DEBUG nova.network.neutron [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.131522] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 517.208172] env[61991]: DEBUG nova.compute.utils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.210330] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.210640] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 517.375986] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.446479] env[61991]: DEBUG nova.policy [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4274fc12414f4e9c82263992b2b2514f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b9e6bab5d0b4c589ad01dbacae1655a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.456591] env[61991]: DEBUG nova.compute.manager [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Received event network-changed-d92d491b-f4d1-494c-aa31-804bf04a46bc {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 517.456787] env[61991]: DEBUG nova.compute.manager [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Refreshing instance network info cache due to event network-changed-d92d491b-f4d1-494c-aa31-804bf04a46bc. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 517.456972] env[61991]: DEBUG oslo_concurrency.lockutils [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] Acquiring lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.577632] env[61991]: INFO nova.compute.manager [-] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Took 1.08 seconds to deallocate network for instance. [ 517.582380] env[61991]: DEBUG nova.compute.claims [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 517.582380] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.629026] env[61991]: DEBUG oslo_concurrency.lockutils [req-d7f9314b-69e5-4887-a167-a97e23018240 req-912c2abb-cf25-4706-9ce8-8b6615f2abce service nova] Releasing lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.711500] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 517.879936] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Releasing lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.881242] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 517.881242] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 517.884304] env[61991]: DEBUG oslo_concurrency.lockutils [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] Acquired lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.887028] env[61991]: DEBUG nova.network.neutron [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Refreshing network info cache for port d92d491b-f4d1-494c-aa31-804bf04a46bc {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 517.887028] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6383989e-2dc4-4300-8d7e-86b62492ccaa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.896311] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b661ea-4e56-450b-ba71-0511e7033bf5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.950874] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cdecebd4-5075-414d-bce5-7bee66c68cba could not be found. [ 517.951969] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 517.952203] env[61991]: INFO nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Took 0.07 seconds to destroy the instance on the hypervisor. [ 517.956629] env[61991]: DEBUG oslo.service.loopingcall [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 517.956629] env[61991]: DEBUG nova.compute.manager [-] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 517.956629] env[61991]: DEBUG nova.network.neutron [-] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 517.958018] env[61991]: INFO nova.scheduler.client.report [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted allocations for instance 8aaa3fdb-a467-42cd-bf27-d877d79eb4ea [ 518.042049] env[61991]: DEBUG nova.network.neutron [-] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 518.114739] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d1da5f-66d0-4cf0-821a-4f2f08e7958f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.131330] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1405fb45-532a-4f87-a106-7e8fa77fbcfb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.167033] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1e1050-ebea-4498-89b0-33ec8518e4f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.179022] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d1df19-23d9-4f3a-a945-5a88337cafa0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.192114] env[61991]: DEBUG nova.compute.provider_tree [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.197616] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "4ea6d766-cf90-4248-a25f-b3763f1cc7ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.197826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "4ea6d766-cf90-4248-a25f-b3763f1cc7ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.453347] env[61991]: DEBUG nova.network.neutron [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 518.478474] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ee8e9bc5-5f66-4dca-adab-b0d732529d38 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "8aaa3fdb-a467-42cd-bf27-d877d79eb4ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.104s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.552490] env[61991]: DEBUG nova.network.neutron [-] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.698784] env[61991]: DEBUG nova.scheduler.client.report [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 518.722102] env[61991]: DEBUG nova.network.neutron [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.725085] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 518.759374] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 518.759521] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 518.759575] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.759749] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 518.759894] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.760278] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 518.760708] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 518.761147] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 518.761147] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 518.761147] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 518.761495] env[61991]: DEBUG nova.virt.hardware [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 518.762681] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c45c58-4921-4934-8381-3f77046ece09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.772152] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ab7d6f-5638-4261-9bf8-6f65a6924467 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.943581] env[61991]: ERROR nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 518.943581] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.943581] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 518.943581] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 518.943581] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.943581] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.943581] env[61991]: ERROR nova.compute.manager raise self.value [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 518.943581] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 518.943581] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.943581] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 518.944908] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.944908] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 518.944908] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 518.944908] env[61991]: ERROR nova.compute.manager [ 518.944908] env[61991]: Traceback (most recent call last): [ 518.944908] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 518.944908] env[61991]: listener.cb(fileno) [ 518.944908] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.944908] env[61991]: result = function(*args, **kwargs) [ 518.944908] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 518.944908] env[61991]: return func(*args, **kwargs) [ 518.944908] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.944908] env[61991]: raise e [ 518.944908] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.944908] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 518.944908] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 518.944908] env[61991]: created_port_ids = self._update_ports_for_instance( [ 518.944908] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 518.944908] env[61991]: with excutils.save_and_reraise_exception(): [ 518.944908] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.944908] env[61991]: self.force_reraise() [ 518.944908] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.944908] env[61991]: raise self.value [ 518.944908] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 518.944908] env[61991]: updated_port = self._update_port( [ 518.944908] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.944908] env[61991]: _ensure_no_port_binding_failure(port) [ 518.944908] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.944908] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 518.945680] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 518.945680] env[61991]: Removing descriptor: 18 [ 518.945680] env[61991]: ERROR nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Traceback (most recent call last): [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] yield resources [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self.driver.spawn(context, instance, image_meta, [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self._vmops.spawn(context, instance, image_meta, injected_files, [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 518.945680] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] vm_ref = self.build_virtual_machine(instance, [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] vif_infos = vmwarevif.get_vif_info(self._session, [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] for vif in network_info: [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return self._sync_wrapper(fn, *args, **kwargs) [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self.wait() [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self[:] = self._gt.wait() [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return self._exit_event.wait() [ 518.945999] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] result = hub.switch() [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return self.greenlet.switch() [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] result = function(*args, **kwargs) [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return func(*args, **kwargs) [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] raise e [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] nwinfo = self.network_api.allocate_for_instance( [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 518.946355] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] created_port_ids = self._update_ports_for_instance( [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] with excutils.save_and_reraise_exception(): [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self.force_reraise() [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] raise self.value [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] updated_port = self._update_port( [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] _ensure_no_port_binding_failure(port) [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.946704] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] raise exception.PortBindingFailed(port_id=port['id']) [ 518.947043] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 518.947043] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] [ 518.947043] env[61991]: INFO nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Terminating instance [ 518.952021] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquiring lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.952021] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquired lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.952021] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 518.985401] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.056487] env[61991]: INFO nova.compute.manager [-] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Took 1.10 seconds to deallocate network for instance. [ 519.059487] env[61991]: DEBUG nova.compute.claims [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 519.059688] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.074555] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Successfully created port: d1a03935-843c-419e-8312-4f340246b759 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 519.107163] env[61991]: DEBUG nova.compute.manager [req-be4034c8-a964-4b73-8300-28c2f23ed9e9 req-644caff9-ee99-4789-b394-99587c120991 service nova] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Received event network-vif-deleted-d8d7a50c-23e7-422e-9171-8acd4d4e9433 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 519.208965] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.209885] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 519.214985] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.763s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.219776] env[61991]: INFO nova.compute.claims [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.233330] env[61991]: DEBUG oslo_concurrency.lockutils [req-d2abc8b5-4861-43e5-aeb4-524a4f604e69 req-2ae11cd8-d28d-47a8-8bd3-3172ac916c55 service nova] Releasing lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.515154] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.517464] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 519.734724] env[61991]: DEBUG nova.compute.utils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 519.739764] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 519.739764] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 519.746444] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.102346] env[61991]: DEBUG nova.policy [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5bbd58fac9964a7290178ce55775de80', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfd699ed289e4985b618b805236d6ea3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.258023] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 520.261418] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Releasing lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.262271] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 520.263486] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 520.263655] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30903e83-c3a9-48bb-9daf-1a77ffea743c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.276098] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fde1f6-f860-4ad4-805d-b92c9f68e5cf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.307162] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8014b7df-35e5-403c-b0dc-b49602d59060 could not be found. [ 520.307550] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 520.307869] env[61991]: INFO nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Took 0.05 seconds to destroy the instance on the hypervisor. [ 520.308411] env[61991]: DEBUG oslo.service.loopingcall [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.312696] env[61991]: DEBUG nova.compute.manager [-] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.312984] env[61991]: DEBUG nova.network.neutron [-] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 520.364511] env[61991]: DEBUG nova.network.neutron [-] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 520.514652] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef7c3d8-2395-456a-8ca8-7dd8320d2aad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.525791] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767d33a0-f331-4fc3-b80e-c1263fc3f8ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.561086] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f499d787-a435-483e-bbaa-e90d5b67e883 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.569209] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2601e4e8-8ccf-49dd-ba67-b85a5028d357 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.584036] env[61991]: DEBUG nova.compute.provider_tree [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.701642] env[61991]: DEBUG nova.compute.manager [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Received event network-vif-deleted-d92d491b-f4d1-494c-aa31-804bf04a46bc {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 520.701642] env[61991]: DEBUG nova.compute.manager [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Received event network-changed-1b61fba4-f7af-44ad-9dbc-196fd7794dc2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 520.701824] env[61991]: DEBUG nova.compute.manager [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Refreshing instance network info cache due to event network-changed-1b61fba4-f7af-44ad-9dbc-196fd7794dc2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 520.704313] env[61991]: DEBUG oslo_concurrency.lockutils [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] Acquiring lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.704618] env[61991]: DEBUG oslo_concurrency.lockutils [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] Acquired lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.705808] env[61991]: DEBUG nova.network.neutron [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Refreshing network info cache for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 520.866693] env[61991]: DEBUG nova.network.neutron [-] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.910749] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Successfully created port: 5cbe6309-6dec-4f66-9acc-4c4d63483d93 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.088043] env[61991]: DEBUG nova.scheduler.client.report [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 521.269691] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 521.309262] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 521.309712] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 521.309971] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 521.310261] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 521.310522] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 521.311396] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 521.311742] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 521.312212] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 521.312559] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 521.312832] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 521.313142] env[61991]: DEBUG nova.virt.hardware [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 521.315130] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540d9c2d-df57-4679-9e33-ae235d3a5862 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.318643] env[61991]: DEBUG nova.network.neutron [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 521.327382] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec7727f-1f4b-4c73-91e9-0d7394e039bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.371377] env[61991]: INFO nova.compute.manager [-] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Took 1.06 seconds to deallocate network for instance. [ 521.372657] env[61991]: DEBUG nova.compute.claims [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 521.372657] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.507965] env[61991]: DEBUG nova.network.neutron [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.593148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.594092] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.600093] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.900s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.012686] env[61991]: DEBUG oslo_concurrency.lockutils [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] Releasing lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.012945] env[61991]: DEBUG nova.compute.manager [req-faa38994-1123-41c1-a496-286ecd307e4d req-96eae3a2-2dd2-4c47-b50d-2ee3100788a0 service nova] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Received event network-vif-deleted-1b61fba4-f7af-44ad-9dbc-196fd7794dc2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 522.106692] env[61991]: DEBUG nova.compute.utils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.114525] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.114822] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 522.370457] env[61991]: DEBUG nova.policy [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ebb976e0ab4446a694c6cea2b83308c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1adc406259f04e53b4babf1548dfae79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.404996] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c87563-5f15-4eb6-a175-e1e4ca66dc3e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.414787] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2cfe31-d526-412a-8d02-807f2bfa46d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.448988] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242d70c8-2784-468a-8ac2-0d5783f43e40 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.457835] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ebdc1e-fdbb-4e26-831d-1e30d1bd2871 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.471853] env[61991]: DEBUG nova.compute.provider_tree [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.625051] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 522.976401] env[61991]: DEBUG nova.scheduler.client.report [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.481207] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.883s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.481816] env[61991]: ERROR nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Traceback (most recent call last): [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self.driver.spawn(context, instance, image_meta, [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] vm_ref = self.build_virtual_machine(instance, [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.481816] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] for vif in network_info: [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return self._sync_wrapper(fn, *args, **kwargs) [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self.wait() [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self[:] = self._gt.wait() [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return self._exit_event.wait() [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] result = hub.switch() [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.482174] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return self.greenlet.switch() [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] result = function(*args, **kwargs) [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] return func(*args, **kwargs) [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] raise e [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] nwinfo = self.network_api.allocate_for_instance( [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] created_port_ids = self._update_ports_for_instance( [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] with excutils.save_and_reraise_exception(): [ 523.483248] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] self.force_reraise() [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] raise self.value [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] updated_port = self._update_port( [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] _ensure_no_port_binding_failure(port) [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] raise exception.PortBindingFailed(port_id=port['id']) [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] nova.exception.PortBindingFailed: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. [ 523.483758] env[61991]: ERROR nova.compute.manager [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] [ 523.487289] env[61991]: DEBUG nova.compute.utils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 523.487289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.490s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.487289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.487289] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 523.487289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.604s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.489237] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Build of instance f32fee48-de95-4ade-9ed0-092786d3f39c was re-scheduled: Binding failed for port 720150e6-5de7-4610-af0c-8b954d9bd1b1, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 523.489957] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 523.490882] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.491799] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.492991] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 523.497925] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051b386e-43fd-471d-891d-cbeb0f39462c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.514398] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1384fac6-4038-4c2f-b26a-b3646db18a74 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.532746] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c0561d-a0f7-4716-b6e2-59214d919537 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.541139] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2370523d-0940-4daa-8e2e-38e3c4d390f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.577960] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181543MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 523.578154] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.579280] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 523.632917] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 523.679993] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.682189] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.682189] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.682539] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.682539] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.682707] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.683497] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.683497] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.683497] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.683497] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.683666] env[61991]: DEBUG nova.virt.hardware [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.684670] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d757c9e3-c3eb-482d-a364-6500f9fe847e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.696778] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dada57c8-e6e0-4855-8ca8-6d8472082f23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.721875] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquiring lock "9043ea71-0207-42c7-8907-0ae61f210fee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.722929] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Lock "9043ea71-0207-42c7-8907-0ae61f210fee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.732813] env[61991]: ERROR nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 523.732813] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.732813] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.732813] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.732813] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.732813] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.732813] env[61991]: ERROR nova.compute.manager raise self.value [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.732813] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.732813] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.732813] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.733592] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.733592] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.733592] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 523.733592] env[61991]: ERROR nova.compute.manager [ 523.733592] env[61991]: Traceback (most recent call last): [ 523.733592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.733592] env[61991]: listener.cb(fileno) [ 523.733592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.733592] env[61991]: result = function(*args, **kwargs) [ 523.733592] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.733592] env[61991]: return func(*args, **kwargs) [ 523.733592] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.733592] env[61991]: raise e [ 523.733592] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.733592] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 523.733592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.733592] env[61991]: created_port_ids = self._update_ports_for_instance( [ 523.733592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.733592] env[61991]: with excutils.save_and_reraise_exception(): [ 523.733592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.733592] env[61991]: self.force_reraise() [ 523.733592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.733592] env[61991]: raise self.value [ 523.733592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.733592] env[61991]: updated_port = self._update_port( [ 523.733592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.733592] env[61991]: _ensure_no_port_binding_failure(port) [ 523.733592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.733592] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.734746] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 523.734746] env[61991]: Removing descriptor: 19 [ 523.734746] env[61991]: ERROR nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Traceback (most recent call last): [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] yield resources [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self.driver.spawn(context, instance, image_meta, [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.734746] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] vm_ref = self.build_virtual_machine(instance, [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] for vif in network_info: [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return self._sync_wrapper(fn, *args, **kwargs) [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self.wait() [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self[:] = self._gt.wait() [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return self._exit_event.wait() [ 523.735117] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] result = hub.switch() [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return self.greenlet.switch() [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] result = function(*args, **kwargs) [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return func(*args, **kwargs) [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] raise e [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] nwinfo = self.network_api.allocate_for_instance( [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.735695] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] created_port_ids = self._update_ports_for_instance( [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] with excutils.save_and_reraise_exception(): [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self.force_reraise() [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] raise self.value [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] updated_port = self._update_port( [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] _ensure_no_port_binding_failure(port) [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.736042] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] raise exception.PortBindingFailed(port_id=port['id']) [ 523.736379] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 523.736379] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] [ 523.736379] env[61991]: INFO nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Terminating instance [ 523.738254] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquiring lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.738525] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquired lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.738626] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 523.754077] env[61991]: ERROR nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 523.754077] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.754077] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.754077] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.754077] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.754077] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.754077] env[61991]: ERROR nova.compute.manager raise self.value [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.754077] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.754077] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.754077] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.754551] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.754551] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.754551] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 523.754551] env[61991]: ERROR nova.compute.manager [ 523.754551] env[61991]: Traceback (most recent call last): [ 523.754551] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.754551] env[61991]: listener.cb(fileno) [ 523.754551] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.754551] env[61991]: result = function(*args, **kwargs) [ 523.754551] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.754551] env[61991]: return func(*args, **kwargs) [ 523.754551] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.754551] env[61991]: raise e [ 523.754551] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.754551] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 523.754551] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.754551] env[61991]: created_port_ids = self._update_ports_for_instance( [ 523.754551] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.754551] env[61991]: with excutils.save_and_reraise_exception(): [ 523.754551] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.754551] env[61991]: self.force_reraise() [ 523.754551] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.754551] env[61991]: raise self.value [ 523.754551] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.754551] env[61991]: updated_port = self._update_port( [ 523.754551] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.754551] env[61991]: _ensure_no_port_binding_failure(port) [ 523.754551] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.754551] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.755381] env[61991]: nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 523.755381] env[61991]: Removing descriptor: 17 [ 523.756475] env[61991]: ERROR nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Traceback (most recent call last): [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] yield resources [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self.driver.spawn(context, instance, image_meta, [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] vm_ref = self.build_virtual_machine(instance, [ 523.756475] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] for vif in network_info: [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return self._sync_wrapper(fn, *args, **kwargs) [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self.wait() [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self[:] = self._gt.wait() [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return self._exit_event.wait() [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.756803] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] result = hub.switch() [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return self.greenlet.switch() [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] result = function(*args, **kwargs) [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return func(*args, **kwargs) [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] raise e [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] nwinfo = self.network_api.allocate_for_instance( [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] created_port_ids = self._update_ports_for_instance( [ 523.757161] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] with excutils.save_and_reraise_exception(): [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self.force_reraise() [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] raise self.value [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] updated_port = self._update_port( [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] _ensure_no_port_binding_failure(port) [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] raise exception.PortBindingFailed(port_id=port['id']) [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 523.757558] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] [ 523.757880] env[61991]: INFO nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Terminating instance [ 523.758554] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquiring lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.758734] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquired lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.758836] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 523.794834] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Successfully created port: bbaa2da4-9b90-49d4-9e47-901b079b7998 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.861234] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.275066] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93406df8-f01b-4eff-83bf-38015187b221 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.280631] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.288742] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab37c6d3-d840-45bc-a9a7-a9449b474355 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.297979] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.330722] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f121e970-0032-4bd5-be6a-18da1c889c88 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.338657] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef0c927-ba82-438f-a1d6-cee17e91b24f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.353026] env[61991]: DEBUG nova.compute.provider_tree [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.364610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-f32fee48-de95-4ade-9ed0-092786d3f39c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.364610] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 524.364610] env[61991]: DEBUG nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.364610] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 524.413154] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.570530] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.680441] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.788993] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquiring lock "2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.789258] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Lock "2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.856154] env[61991]: DEBUG nova.scheduler.client.report [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.917043] env[61991]: DEBUG nova.network.neutron [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.071843] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Releasing lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.072359] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.072561] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 525.072855] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bba24c1-937c-4760-9bab-78c0a33a03b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.086763] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6727803-3587-4d01-84ff-d83b5af39461 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.124823] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5a9d879c-3220-4a16-b119-2ce6949f273f could not be found. [ 525.125348] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 525.125348] env[61991]: INFO nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 525.125916] env[61991]: DEBUG oslo.service.loopingcall [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.125916] env[61991]: DEBUG nova.compute.manager [-] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.125916] env[61991]: DEBUG nova.network.neutron [-] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 525.183744] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Releasing lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.183925] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.184437] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 525.184722] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d277d276-22eb-40a2-98af-0a4242ccba1f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.188149] env[61991]: DEBUG nova.network.neutron [-] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.195388] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728ad420-bd41-447a-9802-f95497f721c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.206835] env[61991]: DEBUG nova.network.neutron [-] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.225297] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a12782a9-3730-4c2f-826e-d1e68d2c182b could not be found. [ 525.225529] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 525.225818] env[61991]: INFO nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 525.225944] env[61991]: DEBUG oslo.service.loopingcall [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.226187] env[61991]: DEBUG nova.compute.manager [-] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.226283] env[61991]: DEBUG nova.network.neutron [-] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 525.274739] env[61991]: DEBUG nova.network.neutron [-] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.363024] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.363410] env[61991]: ERROR nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Traceback (most recent call last): [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self.driver.spawn(context, instance, image_meta, [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self._vmops.spawn(context, instance, image_meta, injected_files, [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] vm_ref = self.build_virtual_machine(instance, [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] vif_infos = vmwarevif.get_vif_info(self._session, [ 525.363410] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] for vif in network_info: [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return self._sync_wrapper(fn, *args, **kwargs) [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self.wait() [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self[:] = self._gt.wait() [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return self._exit_event.wait() [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] result = hub.switch() [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 525.364377] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return self.greenlet.switch() [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] result = function(*args, **kwargs) [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] return func(*args, **kwargs) [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] raise e [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] nwinfo = self.network_api.allocate_for_instance( [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] created_port_ids = self._update_ports_for_instance( [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] with excutils.save_and_reraise_exception(): [ 525.364747] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] self.force_reraise() [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] raise self.value [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] updated_port = self._update_port( [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] _ensure_no_port_binding_failure(port) [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] raise exception.PortBindingFailed(port_id=port['id']) [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] nova.exception.PortBindingFailed: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. [ 525.365133] env[61991]: ERROR nova.compute.manager [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] [ 525.365418] env[61991]: DEBUG nova.compute.utils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 525.367461] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.742s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.369027] env[61991]: INFO nova.compute.claims [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.371768] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Build of instance ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607 was re-scheduled: Binding failed for port 3edcffb6-8ae2-4953-a164-03156abbf10c, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 525.372450] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 525.372758] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquiring lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.372965] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Acquired lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.373202] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 525.420246] env[61991]: INFO nova.compute.manager [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: f32fee48-de95-4ade-9ed0-092786d3f39c] Took 1.06 seconds to deallocate network for instance. [ 525.714065] env[61991]: INFO nova.compute.manager [-] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Took 0.59 seconds to deallocate network for instance. [ 525.716542] env[61991]: DEBUG nova.compute.claims [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.716706] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.776820] env[61991]: DEBUG nova.network.neutron [-] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.917938] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.091918] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.280737] env[61991]: INFO nova.compute.manager [-] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Took 1.05 seconds to deallocate network for instance. [ 526.283460] env[61991]: DEBUG nova.compute.claims [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.283676] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.457293] env[61991]: INFO nova.scheduler.client.report [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocations for instance f32fee48-de95-4ade-9ed0-092786d3f39c [ 526.597822] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Releasing lock "refresh_cache-ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.598079] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 526.598244] env[61991]: DEBUG nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.598406] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 526.637478] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.689783] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb8ef5c-e9f0-41ab-81a8-385f9b37aa94 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.697782] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a336cc53-7444-48f1-9b03-77012879f328 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.736803] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "3a74ce10-43cb-4c56-a47a-941c8effa14a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.737144] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "3a74ce10-43cb-4c56-a47a-941c8effa14a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.738241] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729916bd-d6c3-425a-895f-8a6237edf946 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.745963] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c831ca-c94f-423c-bc2e-7aa6e12d4a7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.760250] env[61991]: DEBUG nova.compute.provider_tree [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.891554] env[61991]: DEBUG nova.compute.manager [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Received event network-changed-d1a03935-843c-419e-8312-4f340246b759 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.891771] env[61991]: DEBUG nova.compute.manager [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Refreshing instance network info cache due to event network-changed-d1a03935-843c-419e-8312-4f340246b759. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.891982] env[61991]: DEBUG oslo_concurrency.lockutils [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] Acquiring lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.892137] env[61991]: DEBUG oslo_concurrency.lockutils [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] Acquired lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.892350] env[61991]: DEBUG nova.network.neutron [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Refreshing network info cache for port d1a03935-843c-419e-8312-4f340246b759 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 526.967088] env[61991]: DEBUG oslo_concurrency.lockutils [None req-749eade0-9b07-4443-8993-8f35e6e16982 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "f32fee48-de95-4ade-9ed0-092786d3f39c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.408s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.107739] env[61991]: DEBUG nova.compute.manager [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Received event network-changed-5cbe6309-6dec-4f66-9acc-4c4d63483d93 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.107952] env[61991]: DEBUG nova.compute.manager [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Refreshing instance network info cache due to event network-changed-5cbe6309-6dec-4f66-9acc-4c4d63483d93. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.108357] env[61991]: DEBUG oslo_concurrency.lockutils [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] Acquiring lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.108357] env[61991]: DEBUG oslo_concurrency.lockutils [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] Acquired lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.108489] env[61991]: DEBUG nova.network.neutron [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Refreshing network info cache for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 527.141386] env[61991]: DEBUG nova.network.neutron [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.165307] env[61991]: ERROR nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 527.165307] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.165307] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.165307] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.165307] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.165307] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.165307] env[61991]: ERROR nova.compute.manager raise self.value [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.165307] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.165307] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.165307] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.165767] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.165767] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.165767] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 527.165767] env[61991]: ERROR nova.compute.manager [ 527.165767] env[61991]: Traceback (most recent call last): [ 527.165767] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.165767] env[61991]: listener.cb(fileno) [ 527.165767] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.165767] env[61991]: result = function(*args, **kwargs) [ 527.165767] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.165767] env[61991]: return func(*args, **kwargs) [ 527.165767] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.165767] env[61991]: raise e [ 527.165767] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.165767] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 527.165767] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.165767] env[61991]: created_port_ids = self._update_ports_for_instance( [ 527.165767] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.165767] env[61991]: with excutils.save_and_reraise_exception(): [ 527.165767] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.165767] env[61991]: self.force_reraise() [ 527.165767] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.165767] env[61991]: raise self.value [ 527.165767] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.165767] env[61991]: updated_port = self._update_port( [ 527.165767] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.165767] env[61991]: _ensure_no_port_binding_failure(port) [ 527.165767] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.165767] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.166551] env[61991]: nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 527.166551] env[61991]: Removing descriptor: 18 [ 527.166551] env[61991]: ERROR nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Traceback (most recent call last): [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] yield resources [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self.driver.spawn(context, instance, image_meta, [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.166551] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] vm_ref = self.build_virtual_machine(instance, [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] for vif in network_info: [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return self._sync_wrapper(fn, *args, **kwargs) [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self.wait() [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self[:] = self._gt.wait() [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return self._exit_event.wait() [ 527.167128] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] result = hub.switch() [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return self.greenlet.switch() [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] result = function(*args, **kwargs) [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return func(*args, **kwargs) [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] raise e [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] nwinfo = self.network_api.allocate_for_instance( [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.167531] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] created_port_ids = self._update_ports_for_instance( [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] with excutils.save_and_reraise_exception(): [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self.force_reraise() [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] raise self.value [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] updated_port = self._update_port( [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] _ensure_no_port_binding_failure(port) [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.168012] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] raise exception.PortBindingFailed(port_id=port['id']) [ 527.168329] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 527.168329] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] [ 527.168329] env[61991]: INFO nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Terminating instance [ 527.171363] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquiring lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.171517] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquired lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.171680] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 527.263211] env[61991]: DEBUG nova.scheduler.client.report [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.419142] env[61991]: DEBUG nova.network.neutron [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.469558] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.596037] env[61991]: DEBUG nova.network.neutron [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.639345] env[61991]: DEBUG nova.network.neutron [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.643599] env[61991]: INFO nova.compute.manager [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] [instance: ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607] Took 1.04 seconds to deallocate network for instance. [ 527.704417] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.774821] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.775089] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.777620] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.196s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.846923] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.855295] env[61991]: DEBUG nova.network.neutron [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.996683] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.099189] env[61991]: DEBUG oslo_concurrency.lockutils [req-5d55f173-4444-4cff-89ba-32083f7def50 req-32dce55f-e14b-4e93-b522-a9402436dfe6 service nova] Releasing lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.283127] env[61991]: DEBUG nova.compute.utils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.289957] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.289957] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 528.349754] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Releasing lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.350263] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 528.350887] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 528.351071] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b8da4cf-f5ac-4dd9-b7d7-318cd5aeef05 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.359369] env[61991]: DEBUG oslo_concurrency.lockutils [req-a9afef2b-5429-4f04-bc4d-67b2412fabff req-4bafeecd-50ac-44f6-a15e-e0ce6f26bf67 service nova] Releasing lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.364082] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acc0eda-16c3-43f7-91ff-856514670c00 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.389215] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8564e62-162a-4df1-a9ac-8826879cec38 could not be found. [ 528.389467] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 528.389683] env[61991]: INFO nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Took 0.04 seconds to destroy the instance on the hypervisor. [ 528.389893] env[61991]: DEBUG oslo.service.loopingcall [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.394113] env[61991]: DEBUG nova.compute.manager [-] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.394113] env[61991]: DEBUG nova.network.neutron [-] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 528.407937] env[61991]: DEBUG nova.policy [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72acc61cb89d41fdb6fc577da5745c7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da8f4730bb504eb1889a2a1397832456', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.440303] env[61991]: DEBUG nova.network.neutron [-] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.621223] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7b56fe-de21-4ff2-aabe-96edeb918140 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.635766] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e335d9b-045b-4578-9072-f0daa259f46b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.682654] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77dd5ce-b807-42e9-822b-e707d77ddcc4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.690660] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3ee9fb-6eea-4424-876d-754b2b62cac8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.705470] env[61991]: DEBUG nova.compute.provider_tree [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.715713] env[61991]: INFO nova.scheduler.client.report [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Deleted allocations for instance ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607 [ 528.754661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.755446] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.793995] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.943056] env[61991]: DEBUG nova.network.neutron [-] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.071713] env[61991]: DEBUG nova.compute.manager [None req-5b154b2f-580c-4453-95b1-c7257b2e9131 tempest-ServerDiagnosticsV248Test-85672924 tempest-ServerDiagnosticsV248Test-85672924-project-admin] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 529.075442] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c49dbb-791a-48f9-bd25-7813d70b0715 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.083613] env[61991]: INFO nova.compute.manager [None req-5b154b2f-580c-4453-95b1-c7257b2e9131 tempest-ServerDiagnosticsV248Test-85672924 tempest-ServerDiagnosticsV248Test-85672924-project-admin] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Retrieving diagnostics [ 529.084673] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2efdb6-2b30-45cd-8ce6-0ce31fb2c596 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.149024] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Successfully created port: f0d203ff-730f-4c3d-b29e-6a06ffabb7bc {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.207973] env[61991]: DEBUG nova.scheduler.client.report [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.223693] env[61991]: DEBUG oslo_concurrency.lockutils [None req-569e785f-9150-45f0-8744-a79e03382b76 tempest-TenantUsagesTestJSON-1171720725 tempest-TenantUsagesTestJSON-1171720725-project-member] Lock "ffbd2ff2-6ad1-48fd-8b23-5a9ddee5e607" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.955s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.447498] env[61991]: INFO nova.compute.manager [-] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Took 1.05 seconds to deallocate network for instance. [ 529.449993] env[61991]: DEBUG nova.compute.claims [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 529.450195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.722397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.722397] env[61991]: ERROR nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Traceback (most recent call last): [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self.driver.spawn(context, instance, image_meta, [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.722397] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] vm_ref = self.build_virtual_machine(instance, [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] for vif in network_info: [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return self._sync_wrapper(fn, *args, **kwargs) [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self.wait() [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self[:] = self._gt.wait() [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return self._exit_event.wait() [ 529.723164] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] result = hub.switch() [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return self.greenlet.switch() [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] result = function(*args, **kwargs) [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] return func(*args, **kwargs) [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] raise e [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] nwinfo = self.network_api.allocate_for_instance( [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.723578] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] created_port_ids = self._update_ports_for_instance( [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] with excutils.save_and_reraise_exception(): [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] self.force_reraise() [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] raise self.value [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] updated_port = self._update_port( [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] _ensure_no_port_binding_failure(port) [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.723902] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] raise exception.PortBindingFailed(port_id=port['id']) [ 529.724270] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] nova.exception.PortBindingFailed: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. [ 529.724270] env[61991]: ERROR nova.compute.manager [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] [ 529.724270] env[61991]: DEBUG nova.compute.utils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 529.724270] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.664s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.729182] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Build of instance e4e28fbe-9f51-4404-9d0c-c0d80a11cb38 was re-scheduled: Binding failed for port d8d7a50c-23e7-422e-9171-8acd4d4e9433, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 529.729868] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 529.731709] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.731709] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquired lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.731709] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.731930] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 529.810290] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.847618] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.848111] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.848275] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.848325] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.848461] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.848611] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.848821] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.849012] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.849156] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.849316] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.849488] env[61991]: DEBUG nova.virt.hardware [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.850468] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a66584f-90d0-4839-84e6-274ebce3dc61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.860676] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3e7cdf-cf4f-48b8-aed0-e0e6b7e33ea2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.906273] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "4ce0f385-cee1-4afd-ac38-5624623953ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.907223] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "4ce0f385-cee1-4afd-ac38-5624623953ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.264397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.274925] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.400412] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.539102] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc06540-d806-4a74-8dab-858968cc455e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.552033] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc265646-3ac5-4bf3-9976-cf7a737d9acd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.587299] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a03ce2-3d08-4216-b023-735bfd56c24e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.593882] env[61991]: DEBUG nova.compute.manager [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Received event network-vif-deleted-d1a03935-843c-419e-8312-4f340246b759 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.594102] env[61991]: DEBUG nova.compute.manager [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Received event network-changed-bbaa2da4-9b90-49d4-9e47-901b079b7998 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.594266] env[61991]: DEBUG nova.compute.manager [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Refreshing instance network info cache due to event network-changed-bbaa2da4-9b90-49d4-9e47-901b079b7998. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 530.594469] env[61991]: DEBUG oslo_concurrency.lockutils [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] Acquiring lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.594605] env[61991]: DEBUG oslo_concurrency.lockutils [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] Acquired lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.594758] env[61991]: DEBUG nova.network.neutron [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Refreshing network info cache for port bbaa2da4-9b90-49d4-9e47-901b079b7998 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 530.601222] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda1fc76-df25-418a-8970-2ce2fc08e319 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.617238] env[61991]: DEBUG nova.compute.provider_tree [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.671853] env[61991]: ERROR nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 530.671853] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.671853] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 530.671853] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 530.671853] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.671853] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.671853] env[61991]: ERROR nova.compute.manager raise self.value [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 530.671853] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 530.671853] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.671853] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 530.672530] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.672530] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 530.672530] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 530.672530] env[61991]: ERROR nova.compute.manager [ 530.672530] env[61991]: Traceback (most recent call last): [ 530.672530] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 530.672530] env[61991]: listener.cb(fileno) [ 530.672530] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.672530] env[61991]: result = function(*args, **kwargs) [ 530.672530] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.672530] env[61991]: return func(*args, **kwargs) [ 530.672530] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.672530] env[61991]: raise e [ 530.672530] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.672530] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 530.672530] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 530.672530] env[61991]: created_port_ids = self._update_ports_for_instance( [ 530.672530] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 530.672530] env[61991]: with excutils.save_and_reraise_exception(): [ 530.672530] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.672530] env[61991]: self.force_reraise() [ 530.672530] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.672530] env[61991]: raise self.value [ 530.672530] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 530.672530] env[61991]: updated_port = self._update_port( [ 530.672530] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.672530] env[61991]: _ensure_no_port_binding_failure(port) [ 530.672530] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.672530] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 530.673520] env[61991]: nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 530.673520] env[61991]: Removing descriptor: 18 [ 530.673520] env[61991]: ERROR nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Traceback (most recent call last): [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] yield resources [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self.driver.spawn(context, instance, image_meta, [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 530.673520] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] vm_ref = self.build_virtual_machine(instance, [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] vif_infos = vmwarevif.get_vif_info(self._session, [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] for vif in network_info: [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return self._sync_wrapper(fn, *args, **kwargs) [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self.wait() [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self[:] = self._gt.wait() [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return self._exit_event.wait() [ 530.674162] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] result = hub.switch() [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return self.greenlet.switch() [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] result = function(*args, **kwargs) [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return func(*args, **kwargs) [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] raise e [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] nwinfo = self.network_api.allocate_for_instance( [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 530.674522] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] created_port_ids = self._update_ports_for_instance( [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] with excutils.save_and_reraise_exception(): [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self.force_reraise() [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] raise self.value [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] updated_port = self._update_port( [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] _ensure_no_port_binding_failure(port) [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.674887] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] raise exception.PortBindingFailed(port_id=port['id']) [ 530.675376] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 530.675376] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] [ 530.675376] env[61991]: INFO nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Terminating instance [ 530.676261] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquiring lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.676432] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquired lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.676573] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 530.698990] env[61991]: DEBUG nova.compute.manager [req-4a5aa4c5-397e-4b69-a172-8f20d52a9860 req-a254f21d-8755-4b05-9390-353c032a7855 service nova] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Received event network-vif-deleted-5cbe6309-6dec-4f66-9acc-4c4d63483d93 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.903244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Releasing lock "refresh_cache-e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.908633] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 530.908633] env[61991]: DEBUG nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.908633] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 530.924723] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.120630] env[61991]: DEBUG nova.scheduler.client.report [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.124875] env[61991]: DEBUG nova.network.neutron [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.198758] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquiring lock "39ba999e-8740-4f9a-8b08-516e5d80d233" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.199009] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Lock "39ba999e-8740-4f9a-8b08-516e5d80d233" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.204314] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.212949] env[61991]: DEBUG nova.network.neutron [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.260732] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.428247] env[61991]: DEBUG nova.network.neutron [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.518464] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "d978d68a-0709-4a32-bbaf-3bfa006be85c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.518984] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "d978d68a-0709-4a32-bbaf-3bfa006be85c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.519281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "d978d68a-0709-4a32-bbaf-3bfa006be85c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.519483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "d978d68a-0709-4a32-bbaf-3bfa006be85c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.519708] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "d978d68a-0709-4a32-bbaf-3bfa006be85c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.522290] env[61991]: INFO nova.compute.manager [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Terminating instance [ 531.524337] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "refresh_cache-d978d68a-0709-4a32-bbaf-3bfa006be85c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.524423] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquired lock "refresh_cache-d978d68a-0709-4a32-bbaf-3bfa006be85c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.524596] env[61991]: DEBUG nova.network.neutron [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 531.627923] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.628608] env[61991]: ERROR nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Traceback (most recent call last): [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self.driver.spawn(context, instance, image_meta, [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] vm_ref = self.build_virtual_machine(instance, [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] vif_infos = vmwarevif.get_vif_info(self._session, [ 531.628608] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] for vif in network_info: [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return self._sync_wrapper(fn, *args, **kwargs) [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self.wait() [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self[:] = self._gt.wait() [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return self._exit_event.wait() [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] result = hub.switch() [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 531.629018] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return self.greenlet.switch() [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] result = function(*args, **kwargs) [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] return func(*args, **kwargs) [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] raise e [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] nwinfo = self.network_api.allocate_for_instance( [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] created_port_ids = self._update_ports_for_instance( [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] with excutils.save_and_reraise_exception(): [ 531.629379] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] self.force_reraise() [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] raise self.value [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] updated_port = self._update_port( [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] _ensure_no_port_binding_failure(port) [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] raise exception.PortBindingFailed(port_id=port['id']) [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] nova.exception.PortBindingFailed: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. [ 531.629715] env[61991]: ERROR nova.compute.manager [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] [ 531.630059] env[61991]: DEBUG nova.compute.utils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 531.630636] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.116s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.632552] env[61991]: INFO nova.compute.claims [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.640521] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Build of instance cdecebd4-5075-414d-bce5-7bee66c68cba was re-scheduled: Binding failed for port d92d491b-f4d1-494c-aa31-804bf04a46bc, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 531.641009] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 531.641441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquiring lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.641441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Acquired lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.641525] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 531.716425] env[61991]: DEBUG oslo_concurrency.lockutils [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] Releasing lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.716425] env[61991]: DEBUG nova.compute.manager [req-7d898c12-7347-4c7d-8a2c-1aff808d38b7 req-9ae7eb91-a79d-4ef1-b4a5-d0e2156fdfd5 service nova] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Received event network-vif-deleted-bbaa2da4-9b90-49d4-9e47-901b079b7998 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.764888] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Releasing lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.768972] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 531.768972] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 531.768972] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e41eb628-3886-444c-bf1d-6ef7ef9650f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.777184] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa08b7e-ca40-47af-b0be-00f4302d0b8d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.805592] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef8274a1-b7e9-4fa2-9969-7afbff66ef9f could not be found. [ 531.806189] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 531.806528] env[61991]: INFO nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 531.806902] env[61991]: DEBUG oslo.service.loopingcall [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.807235] env[61991]: DEBUG nova.compute.manager [-] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 531.807424] env[61991]: DEBUG nova.network.neutron [-] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 531.824755] env[61991]: DEBUG nova.network.neutron [-] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.904315] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquiring lock "69db1562-7169-4ebc-899d-c54ee6e03384" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.904699] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Lock "69db1562-7169-4ebc-899d-c54ee6e03384" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.933846] env[61991]: INFO nova.compute.manager [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: e4e28fbe-9f51-4404-9d0c-c0d80a11cb38] Took 1.03 seconds to deallocate network for instance. [ 532.086800] env[61991]: DEBUG nova.network.neutron [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.176758] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.210241] env[61991]: DEBUG nova.network.neutron [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.272812] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.329051] env[61991]: DEBUG nova.network.neutron [-] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.713024] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Releasing lock "refresh_cache-d978d68a-0709-4a32-bbaf-3bfa006be85c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.713515] env[61991]: DEBUG nova.compute.manager [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.713723] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 532.714646] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037bdc6b-c1f7-459b-b864-812857deee87 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.737219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "b7237462-348e-4e88-b415-a301499fab83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.737550] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "b7237462-348e-4e88-b415-a301499fab83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.738067] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 532.740769] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f26d7161-e5c1-4cd6-b558-a93bd00248ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.751967] env[61991]: DEBUG oslo_vmware.api [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 532.751967] env[61991]: value = "task-1129270" [ 532.751967] env[61991]: _type = "Task" [ 532.751967] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.767817] env[61991]: DEBUG oslo_vmware.api [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129270, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.779368] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Releasing lock "refresh_cache-cdecebd4-5075-414d-bce5-7bee66c68cba" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.779530] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 532.779563] env[61991]: DEBUG nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.779829] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 532.831473] env[61991]: INFO nova.compute.manager [-] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Took 1.02 seconds to deallocate network for instance. [ 532.834476] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.844069] env[61991]: DEBUG nova.compute.claims [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 532.844069] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.973361] env[61991]: INFO nova.scheduler.client.report [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Deleted allocations for instance e4e28fbe-9f51-4404-9d0c-c0d80a11cb38 [ 533.050810] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d7a744-4d11-43a9-8f00-fc6f4e46e4f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.058736] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad67477-fa0a-4c67-9df1-fefd5485d731 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.092703] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03b0cdf-2ae6-466e-9ae1-83864a1ec553 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.100426] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fde5cca-8b55-4058-8532-b3e993d94e8d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.118232] env[61991]: DEBUG nova.compute.provider_tree [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.269039] env[61991]: DEBUG oslo_vmware.api [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129270, 'name': PowerOffVM_Task, 'duration_secs': 0.115703} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.269701] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 533.269874] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 533.270137] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-804f2b7b-083b-49d3-88ca-9f305a87eeb5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.307020] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 533.307020] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 533.307020] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleting the datastore file [datastore2] d978d68a-0709-4a32-bbaf-3bfa006be85c {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 533.307020] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8cd2ab6e-e5e3-4b8c-929c-819dc04ba909 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.312480] env[61991]: DEBUG oslo_vmware.api [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for the task: (returnval){ [ 533.312480] env[61991]: value = "task-1129272" [ 533.312480] env[61991]: _type = "Task" [ 533.312480] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.322116] env[61991]: DEBUG oslo_vmware.api [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.342895] env[61991]: DEBUG nova.network.neutron [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.361203] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "22be69a5-6405-4e8c-8cdd-0843847f9228" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.361468] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "22be69a5-6405-4e8c-8cdd-0843847f9228" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.389099] env[61991]: DEBUG nova.compute.manager [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Received event network-changed-f0d203ff-730f-4c3d-b29e-6a06ffabb7bc {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.389099] env[61991]: DEBUG nova.compute.manager [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Refreshing instance network info cache due to event network-changed-f0d203ff-730f-4c3d-b29e-6a06ffabb7bc. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.389099] env[61991]: DEBUG oslo_concurrency.lockutils [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] Acquiring lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.389099] env[61991]: DEBUG oslo_concurrency.lockutils [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] Acquired lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.389099] env[61991]: DEBUG nova.network.neutron [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Refreshing network info cache for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 533.487652] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9c6285e-63bf-4510-89f8-ae7176126ae2 tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "e4e28fbe-9f51-4404-9d0c-c0d80a11cb38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.080s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.625233] env[61991]: DEBUG nova.scheduler.client.report [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.834292] env[61991]: DEBUG oslo_vmware.api [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Task: {'id': task-1129272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09434} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.834292] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 533.834292] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 533.834292] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.834292] env[61991]: INFO nova.compute.manager [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 533.834605] env[61991]: DEBUG oslo.service.loopingcall [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.834605] env[61991]: DEBUG nova.compute.manager [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.834605] env[61991]: DEBUG nova.network.neutron [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 533.850505] env[61991]: INFO nova.compute.manager [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] [instance: cdecebd4-5075-414d-bce5-7bee66c68cba] Took 1.07 seconds to deallocate network for instance. [ 533.865807] env[61991]: DEBUG nova.network.neutron [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.930263] env[61991]: DEBUG nova.network.neutron [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.989727] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.063124] env[61991]: DEBUG nova.network.neutron [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.131976] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.131976] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.139020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.765s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.369607] env[61991]: DEBUG nova.network.neutron [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.518682] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.567544] env[61991]: DEBUG oslo_concurrency.lockutils [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] Releasing lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.568154] env[61991]: DEBUG nova.compute.manager [req-bac1389a-8fc6-45fd-8fbe-29049a7bae9b req-5ce2c351-e6a3-408f-942a-ec6e071b33b9 service nova] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Received event network-vif-deleted-f0d203ff-730f-4c3d-b29e-6a06ffabb7bc {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.650669] env[61991]: DEBUG nova.compute.utils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.652309] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.652423] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 534.731174] env[61991]: DEBUG nova.policy [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1df6553ea390405babb25f618c26f6e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac21b7c53b6d4aeeaea446ee83ad1142', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.871971] env[61991]: INFO nova.compute.manager [-] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Took 1.04 seconds to deallocate network for instance. [ 534.910390] env[61991]: INFO nova.scheduler.client.report [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Deleted allocations for instance cdecebd4-5075-414d-bce5-7bee66c68cba [ 535.034511] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d243a6cb-81bf-4ff5-93f2-8aa22b64794c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.044733] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f23577-433b-47b6-978d-02e0aebd5327 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.083278] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e82f84-0610-4a7b-a619-ed160cad6714 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.090872] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d878f8c-16fe-42da-ac0b-4c8a51235a6b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.106022] env[61991]: DEBUG nova.compute.provider_tree [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.156278] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.259262] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Successfully created port: 740ad088-829b-4275-a471-0939087ab7b7 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.382541] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.430749] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0668cf37-92cc-4fb9-88eb-5e270e4c3b70 tempest-ServerExternalEventsTest-1367004186 tempest-ServerExternalEventsTest-1367004186-project-member] Lock "cdecebd4-5075-414d-bce5-7bee66c68cba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.636s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.609574] env[61991]: DEBUG nova.scheduler.client.report [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.934244] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 536.119265] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.121020] env[61991]: ERROR nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Traceback (most recent call last): [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self.driver.spawn(context, instance, image_meta, [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] vm_ref = self.build_virtual_machine(instance, [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.121020] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] for vif in network_info: [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return self._sync_wrapper(fn, *args, **kwargs) [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self.wait() [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self[:] = self._gt.wait() [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return self._exit_event.wait() [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] result = hub.switch() [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.121400] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return self.greenlet.switch() [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] result = function(*args, **kwargs) [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] return func(*args, **kwargs) [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] raise e [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] nwinfo = self.network_api.allocate_for_instance( [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] created_port_ids = self._update_ports_for_instance( [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] with excutils.save_and_reraise_exception(): [ 536.121764] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] self.force_reraise() [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] raise self.value [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] updated_port = self._update_port( [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] _ensure_no_port_binding_failure(port) [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] raise exception.PortBindingFailed(port_id=port['id']) [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] nova.exception.PortBindingFailed: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. [ 536.122126] env[61991]: ERROR nova.compute.manager [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] [ 536.127036] env[61991]: DEBUG nova.compute.utils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 536.127036] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.547s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.129535] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Build of instance 8014b7df-35e5-403c-b0dc-b49602d59060 was re-scheduled: Binding failed for port 1b61fba4-f7af-44ad-9dbc-196fd7794dc2, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 536.135137] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 536.135137] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquiring lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.135137] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Acquired lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.135137] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.169822] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.200367] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:17:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='323427174',id=31,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-909617860',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.200367] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.200367] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.200566] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.200566] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.200566] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.200566] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.200566] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.200729] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.200729] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.200729] env[61991]: DEBUG nova.virt.hardware [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.204306] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa8168a-c8d2-47d5-86cc-97a772497e8f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.210927] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293534e4-fdfd-4592-9b74-e420eaf9d793 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.466522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.677099] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.803325] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.868782] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "acc99cda-1a6c-4c2b-a2e3-c92d646da723" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.869012] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "acc99cda-1a6c-4c2b-a2e3-c92d646da723" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.180282] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d978d68a-0709-4a32-bbaf-3bfa006be85c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 537.305551] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Releasing lock "refresh_cache-8014b7df-35e5-403c-b0dc-b49602d59060" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.305663] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 537.305838] env[61991]: DEBUG nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.306016] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 537.309974] env[61991]: ERROR nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 537.309974] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.309974] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.309974] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.309974] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.309974] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.309974] env[61991]: ERROR nova.compute.manager raise self.value [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.309974] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.309974] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.309974] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.310457] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.310457] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.310457] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 537.310457] env[61991]: ERROR nova.compute.manager [ 537.310457] env[61991]: Traceback (most recent call last): [ 537.310457] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.310457] env[61991]: listener.cb(fileno) [ 537.310457] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.310457] env[61991]: result = function(*args, **kwargs) [ 537.310457] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.310457] env[61991]: return func(*args, **kwargs) [ 537.310457] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.310457] env[61991]: raise e [ 537.310457] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.310457] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 537.310457] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.310457] env[61991]: created_port_ids = self._update_ports_for_instance( [ 537.310457] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.310457] env[61991]: with excutils.save_and_reraise_exception(): [ 537.310457] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.310457] env[61991]: self.force_reraise() [ 537.310457] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.310457] env[61991]: raise self.value [ 537.310457] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.310457] env[61991]: updated_port = self._update_port( [ 537.310457] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.310457] env[61991]: _ensure_no_port_binding_failure(port) [ 537.310457] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.310457] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.311610] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 537.311610] env[61991]: Removing descriptor: 18 [ 537.312448] env[61991]: ERROR nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Traceback (most recent call last): [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] yield resources [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self.driver.spawn(context, instance, image_meta, [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] vm_ref = self.build_virtual_machine(instance, [ 537.312448] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] for vif in network_info: [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return self._sync_wrapper(fn, *args, **kwargs) [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self.wait() [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self[:] = self._gt.wait() [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return self._exit_event.wait() [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.312847] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] result = hub.switch() [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return self.greenlet.switch() [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] result = function(*args, **kwargs) [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return func(*args, **kwargs) [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] raise e [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] nwinfo = self.network_api.allocate_for_instance( [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] created_port_ids = self._update_ports_for_instance( [ 537.313206] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] with excutils.save_and_reraise_exception(): [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self.force_reraise() [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] raise self.value [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] updated_port = self._update_port( [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] _ensure_no_port_binding_failure(port) [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] raise exception.PortBindingFailed(port_id=port['id']) [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 537.313574] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] [ 537.313983] env[61991]: INFO nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Terminating instance [ 537.316552] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquiring lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.316710] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquired lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.316871] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 537.326500] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.453318] env[61991]: DEBUG nova.compute.manager [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Received event network-changed-740ad088-829b-4275-a471-0939087ab7b7 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 537.453517] env[61991]: DEBUG nova.compute.manager [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Refreshing instance network info cache due to event network-changed-740ad088-829b-4275-a471-0939087ab7b7. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 537.453712] env[61991]: DEBUG oslo_concurrency.lockutils [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] Acquiring lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.687143] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 8014b7df-35e5-403c-b0dc-b49602d59060 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.687447] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance a12782a9-3730-4c2f-826e-d1e68d2c182b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.687447] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 5a9d879c-3220-4a16-b119-2ce6949f273f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.687556] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c8564e62-162a-4df1-a9ac-8826879cec38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.687669] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance ef8274a1-b7e9-4fa2-9969-7afbff66ef9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.687786] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.829182] env[61991]: DEBUG nova.network.neutron [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.863976] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.160572] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.193055] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 38d3d1bf-8607-494a-a991-0bc4de3d7522 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.335209] env[61991]: INFO nova.compute.manager [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] [instance: 8014b7df-35e5-403c-b0dc-b49602d59060] Took 1.03 seconds to deallocate network for instance. [ 538.663234] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Releasing lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.664217] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.664659] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.665478] env[61991]: DEBUG oslo_concurrency.lockutils [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] Acquired lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.665478] env[61991]: DEBUG nova.network.neutron [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Refreshing network info cache for port 740ad088-829b-4275-a471-0939087ab7b7 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.669068] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d583c91-85df-4072-af23-3e28d197dba8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.681329] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f543e4c8-0bc9-4b25-8bf4-66839e1c6503 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.697423] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.710033] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3 could not be found. [ 538.710033] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.710033] env[61991]: INFO nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 538.710033] env[61991]: DEBUG oslo.service.loopingcall [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.710650] env[61991]: DEBUG nova.compute.manager [-] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.710650] env[61991]: DEBUG nova.network.neutron [-] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.758657] env[61991]: DEBUG nova.network.neutron [-] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.204159] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 4ea6d766-cf90-4248-a25f-b3763f1cc7ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.207789] env[61991]: DEBUG nova.network.neutron [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.261551] env[61991]: DEBUG nova.network.neutron [-] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.373780] env[61991]: INFO nova.scheduler.client.report [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Deleted allocations for instance 8014b7df-35e5-403c-b0dc-b49602d59060 [ 539.455139] env[61991]: DEBUG nova.network.neutron [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.712200] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9043ea71-0207-42c7-8907-0ae61f210fee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.768674] env[61991]: INFO nova.compute.manager [-] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Took 1.06 seconds to deallocate network for instance. [ 539.771239] env[61991]: DEBUG nova.compute.claims [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.771491] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.891624] env[61991]: DEBUG oslo_concurrency.lockutils [None req-86d7867a-0d39-4172-8681-29d06d94d998 tempest-ServerDiagnosticsNegativeTest-98497876 tempest-ServerDiagnosticsNegativeTest-98497876-project-member] Lock "8014b7df-35e5-403c-b0dc-b49602d59060" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.510s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.960064] env[61991]: DEBUG oslo_concurrency.lockutils [req-2db6e0e0-ba5a-42b9-8256-9a027fc1c5a1 req-a5799806-c9a6-463f-bedf-f87e6f4b66a2 service nova] Releasing lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.127720] env[61991]: DEBUG nova.compute.manager [req-59984921-fecd-4830-a2ef-0acb9ff01952 req-3915d806-2dab-4d3a-8293-83a853688417 service nova] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Received event network-vif-deleted-740ad088-829b-4275-a471-0939087ab7b7 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.218818] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 540.397251] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.722972] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3a74ce10-43cb-4c56-a47a-941c8effa14a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 540.932143] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.228716] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 541.733804] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 4ce0f385-cee1-4afd-ac38-5624623953ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 542.241537] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 39ba999e-8740-4f9a-8b08-516e5d80d233 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 542.651874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "c2ef2717-4eac-428b-92a8-12a37d7c0719" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.652053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "c2ef2717-4eac-428b-92a8-12a37d7c0719" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.691759] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "3f9a4042-d9e4-4201-8386-18637c39795f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.691969] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3f9a4042-d9e4-4201-8386-18637c39795f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.744680] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 69db1562-7169-4ebc-899d-c54ee6e03384 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.248013] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance b7237462-348e-4e88-b415-a301499fab83 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.754520] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 22be69a5-6405-4e8c-8cdd-0843847f9228 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.760834] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "432ab035-b5b0-4390-bef2-ec2691b4422b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.761061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "432ab035-b5b0-4390-bef2-ec2691b4422b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.262018] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance acc99cda-1a6c-4c2b-a2e3-c92d646da723 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 544.262369] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 544.262495] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 544.632322] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdcc7f0-2d91-496f-8e91-bb2b132809ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.641950] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1217a5-ab4d-47e8-8541-7f7b050c2afa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.675955] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7377d40e-ec52-4153-b8e3-a1c0a8c964e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.684279] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62ab677-2cb5-49dd-9c9f-a945d98aacb7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.701891] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.204090] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.709479] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 545.709733] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.584s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.710820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.993s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.256652] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquiring lock "32736198-a819-4afc-929b-a2c890263059" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.256836] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Lock "32736198-a819-4afc-929b-a2c890263059" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.635543] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cca386-b6be-458c-b157-a1c58b1182aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.644958] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb49ace1-31f5-46a1-8825-edf8dfe35653 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.677691] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ef3e66-4b82-44a9-8cfb-4042b257832a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.687700] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230236f9-97fd-4b36-8897-be8da1ca0c86 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.705653] env[61991]: DEBUG nova.compute.provider_tree [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.212019] env[61991]: DEBUG nova.scheduler.client.report [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.724341] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.724971] env[61991]: ERROR nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Traceback (most recent call last): [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self.driver.spawn(context, instance, image_meta, [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] vm_ref = self.build_virtual_machine(instance, [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.724971] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] for vif in network_info: [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return self._sync_wrapper(fn, *args, **kwargs) [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self.wait() [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self[:] = self._gt.wait() [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return self._exit_event.wait() [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] result = hub.switch() [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.725325] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return self.greenlet.switch() [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] result = function(*args, **kwargs) [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] return func(*args, **kwargs) [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] raise e [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] nwinfo = self.network_api.allocate_for_instance( [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] created_port_ids = self._update_ports_for_instance( [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] with excutils.save_and_reraise_exception(): [ 547.728894] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] self.force_reraise() [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] raise self.value [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] updated_port = self._update_port( [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] _ensure_no_port_binding_failure(port) [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] raise exception.PortBindingFailed(port_id=port['id']) [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] nova.exception.PortBindingFailed: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. [ 547.729388] env[61991]: ERROR nova.compute.manager [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] [ 547.729758] env[61991]: DEBUG nova.compute.utils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.729758] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.443s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.729968] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Build of instance 5a9d879c-3220-4a16-b119-2ce6949f273f was re-scheduled: Binding failed for port 5cbe6309-6dec-4f66-9acc-4c4d63483d93, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.730451] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.730647] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquiring lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.730785] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Acquired lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.730937] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.267745] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.573197] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.655373] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7125649-c734-478a-8326-6daa8a3448aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.663522] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3432c928-f79c-4127-8ef1-74a672d67abb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.698836] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2527e0-9125-4559-ab19-1c90128f0ff5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.709569] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbbacca-2918-45b4-90dd-92f70bd6a408 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.727383] env[61991]: DEBUG nova.compute.provider_tree [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.045017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "f720a778-5f29-4f27-8b84-7bf377f27dc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.045385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "f720a778-5f29-4f27-8b84-7bf377f27dc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.078880] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Releasing lock "refresh_cache-5a9d879c-3220-4a16-b119-2ce6949f273f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.080392] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.080548] env[61991]: DEBUG nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.082174] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 549.119146] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.229850] env[61991]: DEBUG nova.scheduler.client.report [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.626548] env[61991]: DEBUG nova.network.neutron [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.740031] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.740633] env[61991]: ERROR nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Traceback (most recent call last): [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self.driver.spawn(context, instance, image_meta, [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] vm_ref = self.build_virtual_machine(instance, [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.740633] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] for vif in network_info: [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return self._sync_wrapper(fn, *args, **kwargs) [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self.wait() [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self[:] = self._gt.wait() [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return self._exit_event.wait() [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] result = hub.switch() [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.742024] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return self.greenlet.switch() [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] result = function(*args, **kwargs) [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] return func(*args, **kwargs) [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] raise e [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] nwinfo = self.network_api.allocate_for_instance( [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] created_port_ids = self._update_ports_for_instance( [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] with excutils.save_and_reraise_exception(): [ 549.742429] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] self.force_reraise() [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] raise self.value [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] updated_port = self._update_port( [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] _ensure_no_port_binding_failure(port) [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] raise exception.PortBindingFailed(port_id=port['id']) [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] nova.exception.PortBindingFailed: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. [ 549.742843] env[61991]: ERROR nova.compute.manager [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] [ 549.743178] env[61991]: DEBUG nova.compute.utils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 549.743178] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.747s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.746527] env[61991]: INFO nova.compute.claims [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.751206] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Build of instance a12782a9-3730-4c2f-826e-d1e68d2c182b was re-scheduled: Binding failed for port d1a03935-843c-419e-8312-4f340246b759, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 549.753016] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 549.753016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquiring lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.753016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Acquired lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.753968] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 550.128986] env[61991]: INFO nova.compute.manager [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] [instance: 5a9d879c-3220-4a16-b119-2ce6949f273f] Took 1.05 seconds to deallocate network for instance. [ 550.296550] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.501250] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.003213] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Releasing lock "refresh_cache-a12782a9-3730-4c2f-826e-d1e68d2c182b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.004191] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 551.004191] env[61991]: DEBUG nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.004191] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 551.042978] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.173885] env[61991]: INFO nova.scheduler.client.report [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Deleted allocations for instance 5a9d879c-3220-4a16-b119-2ce6949f273f [ 551.227622] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c753360b-253a-4032-afb4-0e7e04b963c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.238243] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037a0d02-97a3-4c52-99cb-0bde15931515 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.281426] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f51ddf5-5986-4b03-bf33-8c030e173d5b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.289619] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d686021b-d513-4991-83cf-8d5dea9d0609 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.302692] env[61991]: DEBUG nova.compute.provider_tree [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.548837] env[61991]: DEBUG nova.network.neutron [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.687116] env[61991]: DEBUG oslo_concurrency.lockutils [None req-287a58e2-024b-4527-be54-eeb539527d40 tempest-VolumesAssistedSnapshotsTest-1642027315 tempest-VolumesAssistedSnapshotsTest-1642027315-project-member] Lock "5a9d879c-3220-4a16-b119-2ce6949f273f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.405s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.805640] env[61991]: DEBUG nova.scheduler.client.report [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.051390] env[61991]: INFO nova.compute.manager [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] [instance: a12782a9-3730-4c2f-826e-d1e68d2c182b] Took 1.05 seconds to deallocate network for instance. [ 552.192609] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.316292] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.316848] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 552.320889] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.871s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.327408] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "931d58eb-3bbc-4d81-b737-95830b342027" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.327623] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "931d58eb-3bbc-4d81-b737-95830b342027" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.719528] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.830681] env[61991]: DEBUG nova.compute.utils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.832408] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 552.832664] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 552.929090] env[61991]: DEBUG nova.policy [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d784c27246f4ba7904d1e8a256fa07d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e287e65d74e4411ba2ed4c38cc712724', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 553.091945] env[61991]: INFO nova.scheduler.client.report [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Deleted allocations for instance a12782a9-3730-4c2f-826e-d1e68d2c182b [ 553.209254] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "878aee87-a1fc-4c0d-b385-44a98f790536" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.209470] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "878aee87-a1fc-4c0d-b385-44a98f790536" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.308318] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4441429e-3b7e-42c0-a597-20d7a11e85f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.316545] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d907d2-80c3-491e-9fa7-247f279aee82 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.349823] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 553.353600] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ad5249-3ff4-45ff-a258-9043cca7c7db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.364200] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862904ca-b0ac-4337-b7ac-b0b9eea17485 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.382357] env[61991]: DEBUG nova.compute.provider_tree [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.527150] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Successfully created port: bb02370c-b6fc-4df0-b2b4-540e80e401e8 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.606380] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32169de7-73d1-4872-9d12-90d10d08cf85 tempest-ServerRescueTestJSONUnderV235-1980585136 tempest-ServerRescueTestJSONUnderV235-1980585136-project-member] Lock "a12782a9-3730-4c2f-826e-d1e68d2c182b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.488s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.887289] env[61991]: DEBUG nova.scheduler.client.report [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.110099] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 554.361042] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 554.388674] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 554.388674] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 554.388903] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.388965] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 554.389104] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.389257] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 554.389462] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 554.389647] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 554.389770] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 554.389966] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 554.390327] env[61991]: DEBUG nova.virt.hardware [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 554.391253] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae9f117-b635-4c9d-a11e-8103c8d9ed04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.394832] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.395459] env[61991]: ERROR nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Traceback (most recent call last): [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self.driver.spawn(context, instance, image_meta, [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] vm_ref = self.build_virtual_machine(instance, [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.395459] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] for vif in network_info: [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return self._sync_wrapper(fn, *args, **kwargs) [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self.wait() [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self[:] = self._gt.wait() [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return self._exit_event.wait() [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] result = hub.switch() [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.395990] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return self.greenlet.switch() [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] result = function(*args, **kwargs) [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] return func(*args, **kwargs) [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] raise e [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] nwinfo = self.network_api.allocate_for_instance( [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] created_port_ids = self._update_ports_for_instance( [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] with excutils.save_and_reraise_exception(): [ 554.396571] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] self.force_reraise() [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] raise self.value [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] updated_port = self._update_port( [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] _ensure_no_port_binding_failure(port) [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] raise exception.PortBindingFailed(port_id=port['id']) [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] nova.exception.PortBindingFailed: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. [ 554.397287] env[61991]: ERROR nova.compute.manager [instance: c8564e62-162a-4df1-a9ac-8826879cec38] [ 554.398321] env[61991]: DEBUG nova.compute.utils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 554.398321] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.134s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.399343] env[61991]: INFO nova.compute.claims [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.402572] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Build of instance c8564e62-162a-4df1-a9ac-8826879cec38 was re-scheduled: Binding failed for port bbaa2da4-9b90-49d4-9e47-901b079b7998, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 554.403118] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 554.403340] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquiring lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.404082] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Acquired lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.404082] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.409565] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b268cc3-da67-4639-b7b5-7f821cd98127 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.658612] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.694259] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquiring lock "84b50fb2-7338-4da7-a1cb-d638c421a36d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.694471] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Lock "84b50fb2-7338-4da7-a1cb-d638c421a36d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.959051] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.970682] env[61991]: ERROR nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 554.970682] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.970682] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.970682] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.970682] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.970682] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.970682] env[61991]: ERROR nova.compute.manager raise self.value [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.970682] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.970682] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.970682] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.971387] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.971387] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.971387] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 554.971387] env[61991]: ERROR nova.compute.manager [ 554.971387] env[61991]: Traceback (most recent call last): [ 554.971387] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.971387] env[61991]: listener.cb(fileno) [ 554.971387] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.971387] env[61991]: result = function(*args, **kwargs) [ 554.971387] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.971387] env[61991]: return func(*args, **kwargs) [ 554.971387] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.971387] env[61991]: raise e [ 554.971387] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.971387] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 554.971387] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.971387] env[61991]: created_port_ids = self._update_ports_for_instance( [ 554.971387] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.971387] env[61991]: with excutils.save_and_reraise_exception(): [ 554.971387] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.971387] env[61991]: self.force_reraise() [ 554.971387] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.971387] env[61991]: raise self.value [ 554.971387] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.971387] env[61991]: updated_port = self._update_port( [ 554.971387] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.971387] env[61991]: _ensure_no_port_binding_failure(port) [ 554.971387] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.971387] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.972889] env[61991]: nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 554.972889] env[61991]: Removing descriptor: 17 [ 554.972889] env[61991]: ERROR nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Traceback (most recent call last): [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] yield resources [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self.driver.spawn(context, instance, image_meta, [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.972889] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] vm_ref = self.build_virtual_machine(instance, [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] for vif in network_info: [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return self._sync_wrapper(fn, *args, **kwargs) [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self.wait() [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self[:] = self._gt.wait() [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return self._exit_event.wait() [ 554.973463] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] result = hub.switch() [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return self.greenlet.switch() [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] result = function(*args, **kwargs) [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return func(*args, **kwargs) [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] raise e [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] nwinfo = self.network_api.allocate_for_instance( [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.974473] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] created_port_ids = self._update_ports_for_instance( [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] with excutils.save_and_reraise_exception(): [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self.force_reraise() [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] raise self.value [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] updated_port = self._update_port( [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] _ensure_no_port_binding_failure(port) [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.975311] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] raise exception.PortBindingFailed(port_id=port['id']) [ 554.976039] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 554.976039] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] [ 554.976039] env[61991]: INFO nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Terminating instance [ 554.976039] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquiring lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.976039] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquired lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.976039] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 555.045014] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "2866aec9-56fd-4e77-b651-9d85783fb94f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.045536] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.059170] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.192817] env[61991]: DEBUG nova.compute.manager [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Received event network-changed-bb02370c-b6fc-4df0-b2b4-540e80e401e8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.192817] env[61991]: DEBUG nova.compute.manager [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Refreshing instance network info cache due to event network-changed-bb02370c-b6fc-4df0-b2b4-540e80e401e8. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 555.192817] env[61991]: DEBUG oslo_concurrency.lockutils [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] Acquiring lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.510840] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.561832] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Releasing lock "refresh_cache-c8564e62-162a-4df1-a9ac-8826879cec38" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.562076] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 555.562371] env[61991]: DEBUG nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.562452] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.595203] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.705552] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.824711] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ef06fa-6605-4670-9359-7c8a36a0ca35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.832957] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fcb737-e1b8-4a0a-a672-b84393cafed9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.863445] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7e585a-23c5-4c05-b199-d17659566bce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.871475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76da8130-4096-480e-9eac-f02875c3932f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.889048] env[61991]: DEBUG nova.compute.provider_tree [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.098302] env[61991]: DEBUG nova.network.neutron [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.208289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Releasing lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.210653] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.210653] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 556.210653] env[61991]: DEBUG oslo_concurrency.lockutils [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] Acquired lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.210653] env[61991]: DEBUG nova.network.neutron [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Refreshing network info cache for port bb02370c-b6fc-4df0-b2b4-540e80e401e8 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 556.211562] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8dbb911f-ec7d-4573-82a9-e42ead6382d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.226601] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afe8a1a-0a96-4292-9580-95fdccf14abc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.257895] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 38d3d1bf-8607-494a-a991-0bc4de3d7522 could not be found. [ 556.258134] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 556.258919] env[61991]: INFO nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Took 0.05 seconds to destroy the instance on the hypervisor. [ 556.258919] env[61991]: DEBUG oslo.service.loopingcall [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.258919] env[61991]: DEBUG nova.compute.manager [-] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.258919] env[61991]: DEBUG nova.network.neutron [-] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 556.287627] env[61991]: DEBUG nova.network.neutron [-] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.392098] env[61991]: DEBUG nova.scheduler.client.report [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.601687] env[61991]: INFO nova.compute.manager [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] [instance: c8564e62-162a-4df1-a9ac-8826879cec38] Took 1.04 seconds to deallocate network for instance. [ 556.734884] env[61991]: DEBUG nova.network.neutron [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.792813] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.793178] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.793446] env[61991]: DEBUG nova.network.neutron [-] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.823925] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.824662] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.831858] env[61991]: DEBUG nova.network.neutron [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.899798] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.899798] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 556.900788] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.058s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.223733] env[61991]: DEBUG nova.compute.manager [req-8eb2788f-12f0-4308-bbad-7a5acc90b061 req-e1e647b2-bda7-4496-ac8c-685e6fe49b7d service nova] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Received event network-vif-deleted-bb02370c-b6fc-4df0-b2b4-540e80e401e8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.297576] env[61991]: INFO nova.compute.manager [-] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Took 1.04 seconds to deallocate network for instance. [ 557.298334] env[61991]: DEBUG nova.compute.claims [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.298640] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.335457] env[61991]: DEBUG oslo_concurrency.lockutils [req-010be3d3-92f3-4ffa-86f7-9c596915217d req-a619bdba-1ab7-4d3d-862f-84c66763fb84 service nova] Releasing lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.405990] env[61991]: DEBUG nova.compute.utils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.415220] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 557.415423] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 557.488432] env[61991]: DEBUG nova.policy [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5c32855087749a0838cdf7940317b8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '758df06d0c6444aea9a6c4c11dd73bc2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 557.642487] env[61991]: INFO nova.scheduler.client.report [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Deleted allocations for instance c8564e62-162a-4df1-a9ac-8826879cec38 [ 557.876440] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cd4f58-0693-404f-9a9b-b89ae2288c62 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.887020] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86673dc-b046-4592-8b8e-4c4b7df18907 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.915513] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37a19d9-5e05-467f-a885-3f6d6f5dc66b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.918237] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 557.925878] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf247d96-2d29-4edf-ab9d-3df8cff2c663 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.944692] env[61991]: DEBUG nova.compute.provider_tree [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.050098] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Successfully created port: 6733ab24-7f31-4eda-b763-d4552fe0235d {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.152608] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b459ef7-4b29-4e73-861e-f1fbad575bfb tempest-ServerDiagnosticsTest-395221080 tempest-ServerDiagnosticsTest-395221080-project-member] Lock "c8564e62-162a-4df1-a9ac-8826879cec38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.735s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.448560] env[61991]: DEBUG nova.scheduler.client.report [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.654788] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 558.930085] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 558.956154] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.956154] env[61991]: ERROR nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Traceback (most recent call last): [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self.driver.spawn(context, instance, image_meta, [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.956154] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] vm_ref = self.build_virtual_machine(instance, [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] for vif in network_info: [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return self._sync_wrapper(fn, *args, **kwargs) [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self.wait() [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self[:] = self._gt.wait() [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return self._exit_event.wait() [ 558.956653] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] result = hub.switch() [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return self.greenlet.switch() [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] result = function(*args, **kwargs) [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] return func(*args, **kwargs) [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] raise e [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] nwinfo = self.network_api.allocate_for_instance( [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 558.957057] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] created_port_ids = self._update_ports_for_instance( [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] with excutils.save_and_reraise_exception(): [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] self.force_reraise() [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] raise self.value [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] updated_port = self._update_port( [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] _ensure_no_port_binding_failure(port) [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.957443] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] raise exception.PortBindingFailed(port_id=port['id']) [ 558.957802] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] nova.exception.PortBindingFailed: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. [ 558.957802] env[61991]: ERROR nova.compute.manager [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] [ 558.957802] env[61991]: DEBUG nova.compute.utils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 558.958129] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.440s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.960332] env[61991]: INFO nova.compute.claims [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.964502] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Build of instance ef8274a1-b7e9-4fa2-9969-7afbff66ef9f was re-scheduled: Binding failed for port f0d203ff-730f-4c3d-b29e-6a06ffabb7bc, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 558.966182] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 558.966182] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquiring lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.966182] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Acquired lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.966182] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 558.978070] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.978070] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.978266] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.978386] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.978561] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.978672] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.978872] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.979034] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.979196] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.979351] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.979523] env[61991]: DEBUG nova.virt.hardware [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.980426] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb905028-efdb-448a-bf71-b25a3aa30678 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.991525] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74a1c1b-f88e-40fb-b291-5566c572c0ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.179590] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.384096] env[61991]: DEBUG nova.compute.manager [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Received event network-changed-6733ab24-7f31-4eda-b763-d4552fe0235d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 559.385016] env[61991]: DEBUG nova.compute.manager [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Refreshing instance network info cache due to event network-changed-6733ab24-7f31-4eda-b763-d4552fe0235d. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 559.385016] env[61991]: DEBUG oslo_concurrency.lockutils [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] Acquiring lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.385016] env[61991]: DEBUG oslo_concurrency.lockutils [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] Acquired lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.385016] env[61991]: DEBUG nova.network.neutron [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Refreshing network info cache for port 6733ab24-7f31-4eda-b763-d4552fe0235d {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 559.447158] env[61991]: ERROR nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 559.447158] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.447158] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.447158] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.447158] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.447158] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.447158] env[61991]: ERROR nova.compute.manager raise self.value [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.447158] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.447158] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.447158] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.447830] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.447830] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.447830] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 559.447830] env[61991]: ERROR nova.compute.manager [ 559.447830] env[61991]: Traceback (most recent call last): [ 559.447830] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.447830] env[61991]: listener.cb(fileno) [ 559.447830] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.447830] env[61991]: result = function(*args, **kwargs) [ 559.447830] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.447830] env[61991]: return func(*args, **kwargs) [ 559.447830] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.447830] env[61991]: raise e [ 559.447830] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.447830] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 559.447830] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.447830] env[61991]: created_port_ids = self._update_ports_for_instance( [ 559.447830] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.447830] env[61991]: with excutils.save_and_reraise_exception(): [ 559.447830] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.447830] env[61991]: self.force_reraise() [ 559.447830] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.447830] env[61991]: raise self.value [ 559.447830] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.447830] env[61991]: updated_port = self._update_port( [ 559.447830] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.447830] env[61991]: _ensure_no_port_binding_failure(port) [ 559.447830] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.447830] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.448977] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 559.448977] env[61991]: Removing descriptor: 18 [ 559.448977] env[61991]: ERROR nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Traceback (most recent call last): [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] yield resources [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self.driver.spawn(context, instance, image_meta, [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.448977] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] vm_ref = self.build_virtual_machine(instance, [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] for vif in network_info: [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return self._sync_wrapper(fn, *args, **kwargs) [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self.wait() [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self[:] = self._gt.wait() [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return self._exit_event.wait() [ 559.449436] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] result = hub.switch() [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return self.greenlet.switch() [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] result = function(*args, **kwargs) [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return func(*args, **kwargs) [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] raise e [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] nwinfo = self.network_api.allocate_for_instance( [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.449867] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] created_port_ids = self._update_ports_for_instance( [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] with excutils.save_and_reraise_exception(): [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self.force_reraise() [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] raise self.value [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] updated_port = self._update_port( [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] _ensure_no_port_binding_failure(port) [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.450322] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] raise exception.PortBindingFailed(port_id=port['id']) [ 559.450738] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 559.450738] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] [ 559.450738] env[61991]: INFO nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Terminating instance [ 559.450738] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquiring lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.491138] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.576093] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.922375] env[61991]: DEBUG nova.network.neutron [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.019846] env[61991]: DEBUG nova.network.neutron [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.079541] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Releasing lock "refresh_cache-ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.079793] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 560.079961] env[61991]: DEBUG nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.080142] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.103419] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.405183] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f74f4f3-680f-485f-9e2b-52569e56928d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.414432] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4dad8d-b021-42e1-9932-9541fc7ff804 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.446475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4d162a-abce-4851-995a-46b9dba344be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.454317] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bce3955-b237-4035-9f0d-95cca6380688 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.467818] env[61991]: DEBUG nova.compute.provider_tree [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.525703] env[61991]: DEBUG oslo_concurrency.lockutils [req-c4e6aba7-d08b-417c-9187-800ec43038c4 req-02b2c5bc-5542-4815-8f6d-f7cf35111ffd service nova] Releasing lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.526222] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquired lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.526281] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 560.607410] env[61991]: DEBUG nova.network.neutron [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.974054] env[61991]: DEBUG nova.scheduler.client.report [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.044989] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.110085] env[61991]: INFO nova.compute.manager [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] [instance: ef8274a1-b7e9-4fa2-9969-7afbff66ef9f] Took 1.03 seconds to deallocate network for instance. [ 561.140495] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.408419] env[61991]: DEBUG nova.compute.manager [req-de713dcc-4ce0-467e-b91e-70327a019511 req-9d314410-f5af-4cce-b593-baaa51eaf2a2 service nova] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Received event network-vif-deleted-6733ab24-7f31-4eda-b763-d4552fe0235d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.478750] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.479530] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 561.482343] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.100s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.486261] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.486261] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.018s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.486261] env[61991]: INFO nova.compute.claims [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.518612] env[61991]: INFO nova.scheduler.client.report [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Deleted allocations for instance d978d68a-0709-4a32-bbaf-3bfa006be85c [ 561.643536] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Releasing lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.643536] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.643738] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 561.643943] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ef2cdce-d45c-4791-b69b-413ecc4e84eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.654377] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ec187d-d6c1-4ee0-ab00-1c1a878fcb74 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.683072] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65 could not be found. [ 561.683072] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 561.683072] env[61991]: INFO nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Took 0.04 seconds to destroy the instance on the hypervisor. [ 561.683072] env[61991]: DEBUG oslo.service.loopingcall [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.683072] env[61991]: DEBUG nova.compute.manager [-] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.683072] env[61991]: DEBUG nova.network.neutron [-] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 561.709712] env[61991]: DEBUG nova.network.neutron [-] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.992573] env[61991]: DEBUG nova.compute.utils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.993579] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 561.993753] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 562.032299] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03a51197-cb39-4766-a147-7b2dd8adcbc2 tempest-ServerDiagnosticsV248Test-89879971 tempest-ServerDiagnosticsV248Test-89879971-project-member] Lock "d978d68a-0709-4a32-bbaf-3bfa006be85c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.513s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.154951] env[61991]: INFO nova.scheduler.client.report [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Deleted allocations for instance ef8274a1-b7e9-4fa2-9969-7afbff66ef9f [ 562.184174] env[61991]: DEBUG nova.policy [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2666f4b636d4e4d8682a5876772d381', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '518b363cb2db43c288d25c39a239c9c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.211788] env[61991]: DEBUG nova.network.neutron [-] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.503723] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 562.666397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d9ff3f7-7ffe-45cd-bef7-fa026c52795c tempest-ServersAdminNegativeTestJSON-207055502 tempest-ServersAdminNegativeTestJSON-207055502-project-member] Lock "ef8274a1-b7e9-4fa2-9969-7afbff66ef9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.382s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.714875] env[61991]: INFO nova.compute.manager [-] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Took 1.03 seconds to deallocate network for instance. [ 562.723113] env[61991]: DEBUG nova.compute.claims [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.723306] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.740434] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Successfully created port: 250a68b4-a552-4b12-b6f4-93816196dec8 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.901483] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4648c7-4054-4784-abf8-5a968507e9a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.911264] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8414e5-60cd-4961-af80-ee57f1b9b7ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.947491] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49be85f-64e1-4387-b647-fe23ac0d784f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.955931] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b33319b-5bf9-4fcb-beef-95b054daaf1f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.973045] env[61991]: DEBUG nova.compute.provider_tree [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.170009] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.479137] env[61991]: DEBUG nova.scheduler.client.report [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.516294] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 563.562360] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.562555] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.562832] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.562954] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.563197] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.563395] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.563645] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.563835] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.564053] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.564255] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.564541] env[61991]: DEBUG nova.virt.hardware [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.565671] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f8cab9-8780-4a72-85e8-2e382330713f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.574271] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ed7a90-4cc2-4cda-8860-6377c1195ce0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.701524] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.905986] env[61991]: DEBUG nova.compute.manager [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Received event network-changed-250a68b4-a552-4b12-b6f4-93816196dec8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.905986] env[61991]: DEBUG nova.compute.manager [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Refreshing instance network info cache due to event network-changed-250a68b4-a552-4b12-b6f4-93816196dec8. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 563.905986] env[61991]: DEBUG oslo_concurrency.lockutils [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] Acquiring lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.906578] env[61991]: DEBUG oslo_concurrency.lockutils [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] Acquired lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.907035] env[61991]: DEBUG nova.network.neutron [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Refreshing network info cache for port 250a68b4-a552-4b12-b6f4-93816196dec8 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 563.982606] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.982606] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.985663] env[61991]: ERROR nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 563.985663] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.985663] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.985663] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.985663] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.985663] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.985663] env[61991]: ERROR nova.compute.manager raise self.value [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.985663] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.985663] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.985663] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.986251] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.986251] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.986251] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 563.986251] env[61991]: ERROR nova.compute.manager [ 563.986251] env[61991]: Traceback (most recent call last): [ 563.986462] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.986462] env[61991]: listener.cb(fileno) [ 563.986462] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.986462] env[61991]: result = function(*args, **kwargs) [ 563.986462] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.986462] env[61991]: return func(*args, **kwargs) [ 563.986462] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.986462] env[61991]: raise e [ 563.986462] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.986462] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 563.986462] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.986462] env[61991]: created_port_ids = self._update_ports_for_instance( [ 563.986462] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.986462] env[61991]: with excutils.save_and_reraise_exception(): [ 563.986462] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.986462] env[61991]: self.force_reraise() [ 563.986462] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.986462] env[61991]: raise self.value [ 563.986462] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.986462] env[61991]: updated_port = self._update_port( [ 563.986462] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.986462] env[61991]: _ensure_no_port_binding_failure(port) [ 563.986462] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.986462] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.986462] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 563.987243] env[61991]: Removing descriptor: 18 [ 563.987895] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.216s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.999444] env[61991]: ERROR nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Traceback (most recent call last): [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] yield resources [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self.driver.spawn(context, instance, image_meta, [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] vm_ref = self.build_virtual_machine(instance, [ 563.999444] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] for vif in network_info: [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return self._sync_wrapper(fn, *args, **kwargs) [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self.wait() [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self[:] = self._gt.wait() [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return self._exit_event.wait() [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.001254] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] result = hub.switch() [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return self.greenlet.switch() [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] result = function(*args, **kwargs) [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return func(*args, **kwargs) [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] raise e [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] nwinfo = self.network_api.allocate_for_instance( [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] created_port_ids = self._update_ports_for_instance( [ 564.001851] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] with excutils.save_and_reraise_exception(): [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self.force_reraise() [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] raise self.value [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] updated_port = self._update_port( [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] _ensure_no_port_binding_failure(port) [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] raise exception.PortBindingFailed(port_id=port['id']) [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 564.002330] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] [ 564.002800] env[61991]: INFO nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Terminating instance [ 564.004512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.430078] env[61991]: DEBUG nova.network.neutron [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.509968] env[61991]: DEBUG nova.compute.utils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.517048] env[61991]: DEBUG nova.network.neutron [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.517048] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 564.517048] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 564.603332] env[61991]: DEBUG nova.policy [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '86f82e9e03b844aabb98e599ed081c6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aae54b3b7d1048f5bb8aa55c45af22f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.907633] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7895d09-40dd-426a-9fec-dec1d3e0d435 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.922109] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a61686f-0e55-4305-af55-8d44d5a3e4fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.954908] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Successfully created port: 2bd1b153-8242-407a-9733-d5565cb6c677 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.957745] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389f0d67-9e24-4197-9d7f-6140a5fdd00c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.965695] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3aeaf51-8e6f-4ae9-859e-94b8f91c5eae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.984314] env[61991]: DEBUG nova.compute.provider_tree [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.015768] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.023250] env[61991]: DEBUG oslo_concurrency.lockutils [req-5d70bbe9-1e80-4f50-b7ec-bfcff856c384 req-99b93f30-940d-44fe-b679-e6f5eeccd094 service nova] Releasing lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.023580] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.023799] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.491019] env[61991]: DEBUG nova.scheduler.client.report [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.562356] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.694527] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.996051] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.996467] env[61991]: ERROR nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Traceback (most recent call last): [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self.driver.spawn(context, instance, image_meta, [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] vm_ref = self.build_virtual_machine(instance, [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.996467] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] for vif in network_info: [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return self._sync_wrapper(fn, *args, **kwargs) [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self.wait() [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self[:] = self._gt.wait() [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return self._exit_event.wait() [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] result = hub.switch() [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.996825] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return self.greenlet.switch() [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] result = function(*args, **kwargs) [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] return func(*args, **kwargs) [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] raise e [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] nwinfo = self.network_api.allocate_for_instance( [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] created_port_ids = self._update_ports_for_instance( [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] with excutils.save_and_reraise_exception(): [ 565.997408] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] self.force_reraise() [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] raise self.value [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] updated_port = self._update_port( [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] _ensure_no_port_binding_failure(port) [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] raise exception.PortBindingFailed(port_id=port['id']) [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] nova.exception.PortBindingFailed: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. [ 565.997903] env[61991]: ERROR nova.compute.manager [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] [ 565.998679] env[61991]: DEBUG nova.compute.utils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.999058] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Build of instance 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3 was re-scheduled: Binding failed for port 740ad088-829b-4275-a471-0939087ab7b7, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 565.999533] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 565.999847] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquiring lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.000082] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Acquired lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.000315] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.002119] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.071s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.007397] env[61991]: INFO nova.compute.claims [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.029036] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.057661] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.057910] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.058080] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.058274] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.058420] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.058610] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.058759] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.058917] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.062967] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.063245] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.063405] env[61991]: DEBUG nova.virt.hardware [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.064713] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd24f75-b1f9-4235-8f7e-f7b8a3554be2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.075742] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3af2a6-2779-4a34-9bd1-3069d46e5f6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.168409] env[61991]: DEBUG nova.compute.manager [req-eca80f99-e442-4f6d-86a7-509091b85ffa req-9add59c2-53f6-494f-9ffa-c41f743405c8 service nova] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Received event network-vif-deleted-250a68b4-a552-4b12-b6f4-93816196dec8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.197522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.199036] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 566.199036] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.199036] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9fda7e3-c06a-41e2-aecc-e62e999eddbe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.212322] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2ac125-aa8c-4fdd-91af-e2aeaff272bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.238322] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4ea6d766-cf90-4248-a25f-b3763f1cc7ee could not be found. [ 566.238322] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 566.238322] env[61991]: INFO nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 566.238322] env[61991]: DEBUG oslo.service.loopingcall [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.238322] env[61991]: DEBUG nova.compute.manager [-] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.238322] env[61991]: DEBUG nova.network.neutron [-] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 566.259124] env[61991]: DEBUG nova.network.neutron [-] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.452240] env[61991]: ERROR nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 566.452240] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.452240] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.452240] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.452240] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.452240] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.452240] env[61991]: ERROR nova.compute.manager raise self.value [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.452240] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 566.452240] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.452240] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 566.452942] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.452942] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 566.452942] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 566.452942] env[61991]: ERROR nova.compute.manager [ 566.452942] env[61991]: Traceback (most recent call last): [ 566.452942] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 566.452942] env[61991]: listener.cb(fileno) [ 566.452942] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.452942] env[61991]: result = function(*args, **kwargs) [ 566.452942] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.452942] env[61991]: return func(*args, **kwargs) [ 566.452942] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.452942] env[61991]: raise e [ 566.452942] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.452942] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 566.452942] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.452942] env[61991]: created_port_ids = self._update_ports_for_instance( [ 566.452942] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.452942] env[61991]: with excutils.save_and_reraise_exception(): [ 566.452942] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.452942] env[61991]: self.force_reraise() [ 566.452942] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.452942] env[61991]: raise self.value [ 566.452942] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.452942] env[61991]: updated_port = self._update_port( [ 566.452942] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.452942] env[61991]: _ensure_no_port_binding_failure(port) [ 566.452942] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.452942] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 566.454275] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 566.454275] env[61991]: Removing descriptor: 18 [ 566.454275] env[61991]: ERROR nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Traceback (most recent call last): [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] yield resources [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self.driver.spawn(context, instance, image_meta, [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.454275] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] vm_ref = self.build_virtual_machine(instance, [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] for vif in network_info: [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return self._sync_wrapper(fn, *args, **kwargs) [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self.wait() [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self[:] = self._gt.wait() [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return self._exit_event.wait() [ 566.454699] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] result = hub.switch() [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return self.greenlet.switch() [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] result = function(*args, **kwargs) [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return func(*args, **kwargs) [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] raise e [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] nwinfo = self.network_api.allocate_for_instance( [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.455150] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] created_port_ids = self._update_ports_for_instance( [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] with excutils.save_and_reraise_exception(): [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self.force_reraise() [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] raise self.value [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] updated_port = self._update_port( [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] _ensure_no_port_binding_failure(port) [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.455585] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] raise exception.PortBindingFailed(port_id=port['id']) [ 566.455999] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 566.455999] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] [ 566.455999] env[61991]: INFO nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Terminating instance [ 566.459017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquiring lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.459017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquired lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.459017] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.527273] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.654575] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.763438] env[61991]: DEBUG nova.network.neutron [-] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.985301] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.108661] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.158018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Releasing lock "refresh_cache-3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.158018] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.158018] env[61991]: DEBUG nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.158018] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.176053] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.268246] env[61991]: INFO nova.compute.manager [-] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Took 1.03 seconds to deallocate network for instance. [ 567.270701] env[61991]: DEBUG nova.compute.claims [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.270945] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.389686] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb22d2d-3269-4cad-b847-04659250aa10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.397656] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f9403c-7487-4a7d-a0e5-2df17718841c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.429475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936ce72b-973d-43cf-b445-d73504729d83 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.436993] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66cd6e5-8e42-4813-8b18-d9a6fa8dcfd2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.451992] env[61991]: DEBUG nova.compute.provider_tree [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.611737] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Releasing lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.612199] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 567.612617] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 567.612963] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89985a0a-f5d3-42c1-962b-7413bd90f47e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.622563] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c09263b-ee40-45ac-a2fb-3b0ea88ca6b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.643727] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9043ea71-0207-42c7-8907-0ae61f210fee could not be found. [ 567.643988] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 567.644188] env[61991]: INFO nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Took 0.03 seconds to destroy the instance on the hypervisor. [ 567.644431] env[61991]: DEBUG oslo.service.loopingcall [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.644667] env[61991]: DEBUG nova.compute.manager [-] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.644762] env[61991]: DEBUG nova.network.neutron [-] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.660364] env[61991]: DEBUG nova.network.neutron [-] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.678842] env[61991]: DEBUG nova.network.neutron [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.956498] env[61991]: DEBUG nova.scheduler.client.report [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.166994] env[61991]: DEBUG nova.network.neutron [-] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.181025] env[61991]: INFO nova.compute.manager [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] [instance: 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3] Took 1.02 seconds to deallocate network for instance. [ 568.334302] env[61991]: DEBUG nova.compute.manager [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Received event network-changed-2bd1b153-8242-407a-9733-d5565cb6c677 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.335036] env[61991]: DEBUG nova.compute.manager [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Refreshing instance network info cache due to event network-changed-2bd1b153-8242-407a-9733-d5565cb6c677. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 568.335123] env[61991]: DEBUG oslo_concurrency.lockutils [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] Acquiring lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.335251] env[61991]: DEBUG oslo_concurrency.lockutils [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] Acquired lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.335412] env[61991]: DEBUG nova.network.neutron [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Refreshing network info cache for port 2bd1b153-8242-407a-9733-d5565cb6c677 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 568.460269] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.460818] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.467112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.744s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.467112] env[61991]: INFO nova.compute.claims [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.669817] env[61991]: INFO nova.compute.manager [-] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Took 1.02 seconds to deallocate network for instance. [ 568.673837] env[61991]: DEBUG nova.compute.claims [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 568.674823] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.861904] env[61991]: DEBUG nova.network.neutron [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.972019] env[61991]: DEBUG nova.compute.utils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.973222] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.973545] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 568.979817] env[61991]: DEBUG nova.network.neutron [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.046138] env[61991]: DEBUG nova.policy [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c168f149a90345f49e20d580a28fa036', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72d39c34d390430182b239baf541dda9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.224291] env[61991]: INFO nova.scheduler.client.report [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Deleted allocations for instance 3a6dcd72-9b50-426c-9d38-2351bf9fa6b3 [ 569.415921] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Successfully created port: 121bf0e8-89de-48b1-8d67-4d960901fbf2 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.483045] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.486358] env[61991]: DEBUG oslo_concurrency.lockutils [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] Releasing lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.486594] env[61991]: DEBUG nova.compute.manager [req-a361224e-85f2-4920-ab36-f594f5486b51 req-30595f6d-1ef9-4ba1-aae5-cbbf82184876 service nova] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Received event network-vif-deleted-2bd1b153-8242-407a-9733-d5565cb6c677 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.733342] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7663481d-83ca-4a60-82a1-4acd3db4469b tempest-ServersWithSpecificFlavorTestJSON-360140471 tempest-ServersWithSpecificFlavorTestJSON-360140471-project-member] Lock "3a6dcd72-9b50-426c-9d38-2351bf9fa6b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.505s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.860510] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e89325-b7d3-466b-9594-7dc65e3ca069 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.868307] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a48a23-f8b3-4e6d-ae88-666464c94554 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.899622] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e20529-4703-4ac3-a7f5-98635316f182 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.906837] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fcdabd8-5456-4ec0-9a6c-3ca60d4947a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.919870] env[61991]: DEBUG nova.compute.provider_tree [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.193930] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "a2606541-ff39-4d03-bd19-d0ee8e035c08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.193930] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.238374] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.426181] env[61991]: DEBUG nova.scheduler.client.report [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.495236] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.517465] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.517713] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.517871] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.518072] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.518217] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.518360] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.518560] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.518712] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.518873] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.519042] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.519217] env[61991]: DEBUG nova.virt.hardware [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.520083] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f03f8d-f4e4-4928-b272-4ce5d434a00e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.529633] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f56669-c172-480b-bed6-c7cc34835e1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.544212] env[61991]: ERROR nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 570.544212] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.544212] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 570.544212] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 570.544212] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.544212] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.544212] env[61991]: ERROR nova.compute.manager raise self.value [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 570.544212] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 570.544212] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.544212] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 570.544894] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.544894] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 570.544894] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 570.544894] env[61991]: ERROR nova.compute.manager [ 570.544894] env[61991]: Traceback (most recent call last): [ 570.544894] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 570.544894] env[61991]: listener.cb(fileno) [ 570.544894] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.544894] env[61991]: result = function(*args, **kwargs) [ 570.544894] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.544894] env[61991]: return func(*args, **kwargs) [ 570.544894] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.544894] env[61991]: raise e [ 570.544894] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.544894] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 570.544894] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 570.544894] env[61991]: created_port_ids = self._update_ports_for_instance( [ 570.544894] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 570.544894] env[61991]: with excutils.save_and_reraise_exception(): [ 570.544894] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.544894] env[61991]: self.force_reraise() [ 570.544894] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.544894] env[61991]: raise self.value [ 570.544894] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 570.544894] env[61991]: updated_port = self._update_port( [ 570.544894] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.544894] env[61991]: _ensure_no_port_binding_failure(port) [ 570.544894] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.544894] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 570.545912] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 570.545912] env[61991]: Removing descriptor: 17 [ 570.545912] env[61991]: ERROR nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Traceback (most recent call last): [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] yield resources [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self.driver.spawn(context, instance, image_meta, [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.545912] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] vm_ref = self.build_virtual_machine(instance, [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] for vif in network_info: [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return self._sync_wrapper(fn, *args, **kwargs) [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self.wait() [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self[:] = self._gt.wait() [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return self._exit_event.wait() [ 570.546334] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] result = hub.switch() [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return self.greenlet.switch() [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] result = function(*args, **kwargs) [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return func(*args, **kwargs) [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] raise e [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] nwinfo = self.network_api.allocate_for_instance( [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 570.546935] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] created_port_ids = self._update_ports_for_instance( [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] with excutils.save_and_reraise_exception(): [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self.force_reraise() [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] raise self.value [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] updated_port = self._update_port( [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] _ensure_no_port_binding_failure(port) [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.547501] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] raise exception.PortBindingFailed(port_id=port['id']) [ 570.547924] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 570.547924] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] [ 570.547924] env[61991]: INFO nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Terminating instance [ 570.547924] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquiring lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.547924] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquired lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.547924] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 570.641078] env[61991]: DEBUG nova.compute.manager [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Received event network-changed-121bf0e8-89de-48b1-8d67-4d960901fbf2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 570.641078] env[61991]: DEBUG nova.compute.manager [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Refreshing instance network info cache due to event network-changed-121bf0e8-89de-48b1-8d67-4d960901fbf2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 570.641078] env[61991]: DEBUG oslo_concurrency.lockutils [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] Acquiring lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.757400] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.931203] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.931722] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.939180] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.281s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.941061] env[61991]: INFO nova.compute.claims [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 571.064598] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.161765] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.440539] env[61991]: DEBUG nova.compute.utils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.441968] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 571.442840] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 571.505242] env[61991]: DEBUG nova.policy [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69e4ad858d634a5789edf63ccd01faac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8446aad15ca847c19507c1b44b64b304', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.666836] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Releasing lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.667321] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.667474] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.667791] env[61991]: DEBUG oslo_concurrency.lockutils [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] Acquired lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.667948] env[61991]: DEBUG nova.network.neutron [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Refreshing network info cache for port 121bf0e8-89de-48b1-8d67-4d960901fbf2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 571.671528] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf6d90cb-ca68-47e5-aa7b-947f926092e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.680831] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4469caac-4cbf-4f11-820f-4b4dd752d4a4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.702578] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300 could not be found. [ 571.702806] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.702982] env[61991]: INFO nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Took 0.04 seconds to destroy the instance on the hypervisor. [ 571.703248] env[61991]: DEBUG oslo.service.loopingcall [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.703460] env[61991]: DEBUG nova.compute.manager [-] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.703547] env[61991]: DEBUG nova.network.neutron [-] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.722586] env[61991]: DEBUG nova.network.neutron [-] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.848247] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Successfully created port: 7a72e257-b290-4f08-965e-4491625c0376 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.949022] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 572.208771] env[61991]: DEBUG nova.network.neutron [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.226176] env[61991]: DEBUG nova.network.neutron [-] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.322953] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Successfully created port: 8101f64f-82f2-4ee4-9f80-30bada2de873 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.382596] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb88e232-dfb7-4528-85a0-0cd76b7222bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.390417] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfae440b-de4e-4654-949f-746d4f9f3163 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.420303] env[61991]: DEBUG nova.network.neutron [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.422278] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d202bf4f-a518-466f-b735-534b92ec4d71 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.429213] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a280160-cafe-4cac-a292-405d721243a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.443157] env[61991]: DEBUG nova.compute.provider_tree [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.727521] env[61991]: INFO nova.compute.manager [-] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Took 1.02 seconds to deallocate network for instance. [ 572.732202] env[61991]: DEBUG nova.compute.claims [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.732202] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.739633] env[61991]: DEBUG nova.compute.manager [req-73e544f0-65f5-42f8-b3db-192c9ab3a702 req-9d385ce7-44e1-4256-81c1-7146d47b6c01 service nova] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Received event network-vif-deleted-121bf0e8-89de-48b1-8d67-4d960901fbf2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.925634] env[61991]: DEBUG oslo_concurrency.lockutils [req-cc8638cb-96b1-412d-a878-d46f6dc11097 req-934b942f-c244-43ca-bce5-bf254cc5830c service nova] Releasing lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.949124] env[61991]: DEBUG nova.scheduler.client.report [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.957834] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.969153] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Successfully created port: 594004b9-d5d1-44e1-858b-a7e688f98bf5 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.995717] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.995959] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.996171] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.996368] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.996491] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.996634] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.996838] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.996993] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.997232] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.997395] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.997683] env[61991]: DEBUG nova.virt.hardware [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.998950] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614daab1-d241-40c6-8139-b0d441f23eb7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.008591] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30ecb01-0d6e-470e-ae72-aed33202aa04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.460016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.460016] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 573.460844] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.162s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.966009] env[61991]: DEBUG nova.compute.utils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.975100] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.975100] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 574.079435] env[61991]: DEBUG nova.policy [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a385cf261940a19eb17046958920ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3015571d285418aae0fca50b563394e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 574.394785] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a522a0-cbd2-43da-995c-a13c1f1f3d59 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.404124] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca2e6dc-3c74-4c47-8a1a-ae22e591c80f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.434202] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c49453-1800-41ba-9c3b-bb5c80157b94 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.441996] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2204a6-7dc2-4031-a3f6-76c364232733 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.456775] env[61991]: DEBUG nova.compute.provider_tree [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.474877] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.858940] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Successfully created port: 8db2d35c-9b14-4476-8e29-b5a83c9b70e0 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.959736] env[61991]: DEBUG nova.scheduler.client.report [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.295406] env[61991]: DEBUG nova.compute.manager [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Received event network-changed-7a72e257-b290-4f08-965e-4491625c0376 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.295406] env[61991]: DEBUG nova.compute.manager [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Refreshing instance network info cache due to event network-changed-7a72e257-b290-4f08-965e-4491625c0376. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 575.300381] env[61991]: DEBUG oslo_concurrency.lockutils [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] Acquiring lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.300381] env[61991]: DEBUG oslo_concurrency.lockutils [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] Acquired lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.300381] env[61991]: DEBUG nova.network.neutron [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Refreshing network info cache for port 7a72e257-b290-4f08-965e-4491625c0376 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 575.473096] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.473096] env[61991]: ERROR nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Traceback (most recent call last): [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self.driver.spawn(context, instance, image_meta, [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.473096] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] vm_ref = self.build_virtual_machine(instance, [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] for vif in network_info: [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return self._sync_wrapper(fn, *args, **kwargs) [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self.wait() [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self[:] = self._gt.wait() [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return self._exit_event.wait() [ 575.473623] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] result = hub.switch() [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return self.greenlet.switch() [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] result = function(*args, **kwargs) [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] return func(*args, **kwargs) [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] raise e [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] nwinfo = self.network_api.allocate_for_instance( [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.474172] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] created_port_ids = self._update_ports_for_instance( [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] with excutils.save_and_reraise_exception(): [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] self.force_reraise() [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] raise self.value [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] updated_port = self._update_port( [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] _ensure_no_port_binding_failure(port) [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.474480] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] raise exception.PortBindingFailed(port_id=port['id']) [ 575.474776] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] nova.exception.PortBindingFailed: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. [ 575.474776] env[61991]: ERROR nova.compute.manager [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] [ 575.474776] env[61991]: DEBUG nova.compute.utils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.476332] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Build of instance 38d3d1bf-8607-494a-a991-0bc4de3d7522 was re-scheduled: Binding failed for port bb02370c-b6fc-4df0-b2b4-540e80e401e8, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.476332] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.476521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquiring lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.476521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Acquired lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.476720] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.477793] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.298s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.480243] env[61991]: INFO nova.compute.claims [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.484787] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 575.525820] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.526097] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.526329] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.526683] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.526775] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.527291] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.527524] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.527697] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.528162] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.528162] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.528272] env[61991]: DEBUG nova.virt.hardware [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.529665] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4c07f8-4ea9-484b-b97a-4c270f070d21 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.542953] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3aa980d-4a27-4341-a565-1d66a2154900 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.562133] env[61991]: ERROR nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 575.562133] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.562133] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.562133] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.562133] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.562133] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.562133] env[61991]: ERROR nova.compute.manager raise self.value [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.562133] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.562133] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.562133] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.562571] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.562571] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.562571] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 575.562571] env[61991]: ERROR nova.compute.manager [ 575.562571] env[61991]: Traceback (most recent call last): [ 575.562571] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.562571] env[61991]: listener.cb(fileno) [ 575.562571] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.562571] env[61991]: result = function(*args, **kwargs) [ 575.562571] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.562571] env[61991]: return func(*args, **kwargs) [ 575.562571] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.562571] env[61991]: raise e [ 575.562571] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.562571] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 575.562571] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.562571] env[61991]: created_port_ids = self._update_ports_for_instance( [ 575.562571] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.562571] env[61991]: with excutils.save_and_reraise_exception(): [ 575.562571] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.562571] env[61991]: self.force_reraise() [ 575.562571] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.562571] env[61991]: raise self.value [ 575.562571] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.562571] env[61991]: updated_port = self._update_port( [ 575.562571] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.562571] env[61991]: _ensure_no_port_binding_failure(port) [ 575.562571] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.562571] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.564048] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 575.564048] env[61991]: Removing descriptor: 17 [ 575.564576] env[61991]: ERROR nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Traceback (most recent call last): [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] yield resources [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self.driver.spawn(context, instance, image_meta, [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] vm_ref = self.build_virtual_machine(instance, [ 575.564576] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] for vif in network_info: [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return self._sync_wrapper(fn, *args, **kwargs) [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self.wait() [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self[:] = self._gt.wait() [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return self._exit_event.wait() [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.565208] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] result = hub.switch() [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return self.greenlet.switch() [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] result = function(*args, **kwargs) [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return func(*args, **kwargs) [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] raise e [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] nwinfo = self.network_api.allocate_for_instance( [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] created_port_ids = self._update_ports_for_instance( [ 575.565681] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] with excutils.save_and_reraise_exception(): [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self.force_reraise() [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] raise self.value [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] updated_port = self._update_port( [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] _ensure_no_port_binding_failure(port) [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] raise exception.PortBindingFailed(port_id=port['id']) [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 575.565954] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] [ 575.566341] env[61991]: INFO nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Terminating instance [ 575.570313] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.850914] env[61991]: DEBUG nova.network.neutron [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.148910] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.191122] env[61991]: DEBUG nova.network.neutron [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.375625] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.694163] env[61991]: DEBUG oslo_concurrency.lockutils [req-8146c7f3-6fbf-4c7b-9ecf-190a80f44802 req-35365f18-c51f-4c90-8659-94c8f21e2a6f service nova] Releasing lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.694942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquired lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.694942] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.754998] env[61991]: ERROR nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 576.754998] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.754998] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.754998] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.754998] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.754998] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.754998] env[61991]: ERROR nova.compute.manager raise self.value [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.754998] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 576.754998] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.754998] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 576.755416] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.755416] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 576.755416] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 576.755416] env[61991]: ERROR nova.compute.manager [ 576.755416] env[61991]: Traceback (most recent call last): [ 576.755416] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 576.755416] env[61991]: listener.cb(fileno) [ 576.755416] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.755416] env[61991]: result = function(*args, **kwargs) [ 576.755416] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.755416] env[61991]: return func(*args, **kwargs) [ 576.755416] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.755416] env[61991]: raise e [ 576.755416] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.755416] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 576.755416] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.755416] env[61991]: created_port_ids = self._update_ports_for_instance( [ 576.755416] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.755416] env[61991]: with excutils.save_and_reraise_exception(): [ 576.755416] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.755416] env[61991]: self.force_reraise() [ 576.755416] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.755416] env[61991]: raise self.value [ 576.755416] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.755416] env[61991]: updated_port = self._update_port( [ 576.755416] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.755416] env[61991]: _ensure_no_port_binding_failure(port) [ 576.755416] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.755416] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 576.756181] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 576.756181] env[61991]: Removing descriptor: 18 [ 576.756181] env[61991]: ERROR nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Traceback (most recent call last): [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] yield resources [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self.driver.spawn(context, instance, image_meta, [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.756181] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] vm_ref = self.build_virtual_machine(instance, [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] for vif in network_info: [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return self._sync_wrapper(fn, *args, **kwargs) [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self.wait() [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self[:] = self._gt.wait() [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return self._exit_event.wait() [ 576.756426] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] result = hub.switch() [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return self.greenlet.switch() [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] result = function(*args, **kwargs) [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return func(*args, **kwargs) [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] raise e [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] nwinfo = self.network_api.allocate_for_instance( [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.756678] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] created_port_ids = self._update_ports_for_instance( [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] with excutils.save_and_reraise_exception(): [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self.force_reraise() [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] raise self.value [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] updated_port = self._update_port( [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] _ensure_no_port_binding_failure(port) [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.756933] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] raise exception.PortBindingFailed(port_id=port['id']) [ 576.757230] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 576.757230] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] [ 576.757230] env[61991]: INFO nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Terminating instance [ 576.763375] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.763375] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.763375] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.881680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Releasing lock "refresh_cache-38d3d1bf-8607-494a-a991-0bc4de3d7522" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.881759] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.881924] env[61991]: DEBUG nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.882110] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.902928] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfc9019-0a9c-45b6-b3d1-8ba1e9ed7f73 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.907875] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.914806] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47561b97-eea1-4426-a835-3c80b0bea959 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.957424] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32357cc9-696a-49fd-8115-ae5d9dc9d2a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.963338] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db7ea6c-74c2-4452-8d7c-47e0e7657b55 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.977252] env[61991]: DEBUG nova.compute.provider_tree [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.269113] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.286192] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.378504] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.412089] env[61991]: DEBUG nova.network.neutron [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.423102] env[61991]: DEBUG nova.compute.manager [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Received event network-vif-deleted-7a72e257-b290-4f08-965e-4491625c0376 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.423311] env[61991]: DEBUG nova.compute.manager [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Received event network-changed-8db2d35c-9b14-4476-8e29-b5a83c9b70e0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.423975] env[61991]: DEBUG nova.compute.manager [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Refreshing instance network info cache due to event network-changed-8db2d35c-9b14-4476-8e29-b5a83c9b70e0. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 577.423975] env[61991]: DEBUG oslo_concurrency.lockutils [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] Acquiring lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.483530] env[61991]: DEBUG nova.scheduler.client.report [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.531771] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.884463] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.884463] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.884463] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.884463] env[61991]: DEBUG oslo_concurrency.lockutils [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] Acquired lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.884463] env[61991]: DEBUG nova.network.neutron [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Refreshing network info cache for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.885436] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66fbf011-e63c-47c9-8120-4e3a5397bb10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.896739] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a046b3-d767-4272-b0f0-b34ec56422fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.925020] env[61991]: INFO nova.compute.manager [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] [instance: 38d3d1bf-8607-494a-a991-0bc4de3d7522] Took 1.04 seconds to deallocate network for instance. [ 577.926310] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea could not be found. [ 577.926310] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.926783] env[61991]: INFO nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 577.926879] env[61991]: DEBUG oslo.service.loopingcall [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.927590] env[61991]: DEBUG nova.compute.manager [-] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.928193] env[61991]: DEBUG nova.network.neutron [-] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.955962] env[61991]: DEBUG nova.network.neutron [-] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.988260] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.988803] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.996477] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.269s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.040034] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Releasing lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.040034] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 578.040034] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 578.040034] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d73cd903-1c87-4c68-9cf0-eb481a27fb34 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.051198] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.051198] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.057018] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c227b4de-da1f-4349-a225-93130d2e4407 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.080300] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3a74ce10-43cb-4c56-a47a-941c8effa14a could not be found. [ 578.080748] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 578.081180] env[61991]: INFO nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 578.081406] env[61991]: DEBUG oslo.service.loopingcall [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.081660] env[61991]: DEBUG nova.compute.manager [-] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.081837] env[61991]: DEBUG nova.network.neutron [-] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.264460] env[61991]: DEBUG nova.network.neutron [-] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.412554] env[61991]: DEBUG nova.network.neutron [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.458435] env[61991]: DEBUG nova.network.neutron [-] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.498402] env[61991]: DEBUG nova.compute.utils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.503684] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.504239] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 578.509175] env[61991]: DEBUG nova.network.neutron [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.603533] env[61991]: DEBUG nova.policy [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8946c3371164af6b2943fddd282f977', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7b25d4745e1419b963e387c8461dd7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.969102] env[61991]: INFO nova.compute.manager [-] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Took 1.04 seconds to deallocate network for instance. [ 578.973571] env[61991]: DEBUG nova.compute.claims [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.973976] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.988559] env[61991]: INFO nova.scheduler.client.report [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Deleted allocations for instance 38d3d1bf-8607-494a-a991-0bc4de3d7522 [ 579.006994] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 579.013546] env[61991]: DEBUG oslo_concurrency.lockutils [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] Releasing lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.014477] env[61991]: DEBUG nova.compute.manager [req-c9b48b36-76e2-4673-aec8-2ec9e35390bc req-ec1d4beb-9898-4a12-967a-ab2419fdb6a3 service nova] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Received event network-vif-deleted-8db2d35c-9b14-4476-8e29-b5a83c9b70e0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.037391] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421fb4d0-c088-43f5-848f-10cb0baf5883 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.049906] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbda3484-69dd-4c5e-be28-e319395dd650 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.099338] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0ffea7-b476-439e-9e57-3c14b50ba314 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.112161] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103c7594-a3da-487e-aefd-612d0ddec1c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.142700] env[61991]: DEBUG nova.compute.provider_tree [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.472597] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Successfully created port: 99edbbbc-52da-476d-8302-b494ffa0c42d {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.501332] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0f7ed031-13be-4cb7-a7a3-56028076acae tempest-InstanceActionsV221TestJSON-1033155644 tempest-InstanceActionsV221TestJSON-1033155644-project-member] Lock "38d3d1bf-8607-494a-a991-0bc4de3d7522" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.039s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.649521] env[61991]: DEBUG nova.scheduler.client.report [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.875487] env[61991]: DEBUG nova.network.neutron [-] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.004470] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.022718] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.059142] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.059521] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.059644] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.059996] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.060565] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.060785] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.061016] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.061170] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.061887] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.061887] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.061887] env[61991]: DEBUG nova.virt.hardware [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.063607] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77154454-cc02-4576-861e-19d4a253b08b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.075024] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7b65d9-58bc-49a4-933b-3e1b7f1cecb3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.155036] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.162s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.155783] env[61991]: ERROR nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Traceback (most recent call last): [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self.driver.spawn(context, instance, image_meta, [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] vm_ref = self.build_virtual_machine(instance, [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.155783] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] for vif in network_info: [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return self._sync_wrapper(fn, *args, **kwargs) [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self.wait() [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self[:] = self._gt.wait() [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return self._exit_event.wait() [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] result = hub.switch() [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.156146] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return self.greenlet.switch() [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] result = function(*args, **kwargs) [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] return func(*args, **kwargs) [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] raise e [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] nwinfo = self.network_api.allocate_for_instance( [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] created_port_ids = self._update_ports_for_instance( [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] with excutils.save_and_reraise_exception(): [ 580.156416] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] self.force_reraise() [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] raise self.value [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] updated_port = self._update_port( [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] _ensure_no_port_binding_failure(port) [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] raise exception.PortBindingFailed(port_id=port['id']) [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] nova.exception.PortBindingFailed: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. [ 580.156752] env[61991]: ERROR nova.compute.manager [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] [ 580.156990] env[61991]: DEBUG nova.compute.utils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.157841] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.457s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.159841] env[61991]: INFO nova.compute.claims [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.163218] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Build of instance 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65 was re-scheduled: Binding failed for port 6733ab24-7f31-4eda-b763-d4552fe0235d, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 580.164660] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 580.165117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquiring lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.165313] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Acquired lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.166917] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.378522] env[61991]: INFO nova.compute.manager [-] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Took 2.30 seconds to deallocate network for instance. [ 580.383586] env[61991]: DEBUG nova.compute.claims [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.383627] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.534391] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.562647] env[61991]: DEBUG nova.compute.manager [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Received event network-changed-99edbbbc-52da-476d-8302-b494ffa0c42d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.562895] env[61991]: DEBUG nova.compute.manager [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Refreshing instance network info cache due to event network-changed-99edbbbc-52da-476d-8302-b494ffa0c42d. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 580.565996] env[61991]: DEBUG oslo_concurrency.lockutils [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] Acquiring lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.566179] env[61991]: DEBUG oslo_concurrency.lockutils [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] Acquired lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.566348] env[61991]: DEBUG nova.network.neutron [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Refreshing network info cache for port 99edbbbc-52da-476d-8302-b494ffa0c42d {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 580.627364] env[61991]: ERROR nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 580.627364] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.627364] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.627364] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.627364] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.627364] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.627364] env[61991]: ERROR nova.compute.manager raise self.value [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.627364] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.627364] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.627364] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.627839] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.627839] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.627839] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 580.627839] env[61991]: ERROR nova.compute.manager [ 580.627839] env[61991]: Traceback (most recent call last): [ 580.627839] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.627839] env[61991]: listener.cb(fileno) [ 580.627839] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.627839] env[61991]: result = function(*args, **kwargs) [ 580.627839] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.627839] env[61991]: return func(*args, **kwargs) [ 580.627839] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.627839] env[61991]: raise e [ 580.627839] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.627839] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 580.627839] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.627839] env[61991]: created_port_ids = self._update_ports_for_instance( [ 580.627839] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.627839] env[61991]: with excutils.save_and_reraise_exception(): [ 580.627839] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.627839] env[61991]: self.force_reraise() [ 580.627839] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.627839] env[61991]: raise self.value [ 580.627839] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.627839] env[61991]: updated_port = self._update_port( [ 580.627839] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.627839] env[61991]: _ensure_no_port_binding_failure(port) [ 580.627839] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.627839] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.628610] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 580.628610] env[61991]: Removing descriptor: 17 [ 580.628610] env[61991]: ERROR nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Traceback (most recent call last): [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] yield resources [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self.driver.spawn(context, instance, image_meta, [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.628610] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] vm_ref = self.build_virtual_machine(instance, [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] for vif in network_info: [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return self._sync_wrapper(fn, *args, **kwargs) [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self.wait() [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self[:] = self._gt.wait() [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return self._exit_event.wait() [ 580.628879] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] result = hub.switch() [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return self.greenlet.switch() [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] result = function(*args, **kwargs) [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return func(*args, **kwargs) [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] raise e [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] nwinfo = self.network_api.allocate_for_instance( [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.629165] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] created_port_ids = self._update_ports_for_instance( [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] with excutils.save_and_reraise_exception(): [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self.force_reraise() [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] raise self.value [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] updated_port = self._update_port( [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] _ensure_no_port_binding_failure(port) [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.629483] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] raise exception.PortBindingFailed(port_id=port['id']) [ 580.629763] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 580.629763] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] [ 580.629763] env[61991]: INFO nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Terminating instance [ 580.631765] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.650140] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "812d3491-0598-4171-b905-5304ebf6ae37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.650395] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "812d3491-0598-4171-b905-5304ebf6ae37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.689938] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.787921] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.089675] env[61991]: DEBUG nova.network.neutron [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.187417] env[61991]: DEBUG nova.network.neutron [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.292673] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Releasing lock "refresh_cache-8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.293031] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.293316] env[61991]: DEBUG nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.293540] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.311221] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.603987] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988c4ebc-3c27-4c28-b6c2-c9597382e0bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.612290] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6aa9942-549f-4ed4-b3ff-013cbc4a0b8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.642109] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9be675-17e1-450d-a352-c47ba685307f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.650516] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ad4fb0-208d-4167-8410-676a3b88caa1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.664948] env[61991]: DEBUG nova.compute.provider_tree [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.696579] env[61991]: DEBUG oslo_concurrency.lockutils [req-1805d015-eaf0-48c1-a031-8a4937b81079 req-ef06b85a-35eb-4e8c-860c-b124aa068d40 service nova] Releasing lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.696960] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.698144] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 581.815798] env[61991]: DEBUG nova.network.neutron [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.168426] env[61991]: DEBUG nova.scheduler.client.report [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.216382] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.321604] env[61991]: INFO nova.compute.manager [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] [instance: 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65] Took 1.03 seconds to deallocate network for instance. [ 582.346169] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.617022] env[61991]: DEBUG nova.compute.manager [req-264df894-3ab4-41e2-b58f-9ee163113d72 req-3d26585b-4a4b-4963-8d23-744c88638b46 service nova] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Received event network-vif-deleted-99edbbbc-52da-476d-8302-b494ffa0c42d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.675724] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.675724] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.678722] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.408s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.849170] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.849615] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.849827] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.850141] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe848451-ac76-4757-8b83-e79505d47d15 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.860894] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b13753e-6d0f-4434-ada0-04aa59be4dbe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.883945] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4ce0f385-cee1-4afd-ac38-5624623953ac could not be found. [ 582.884195] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.884376] env[61991]: INFO nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Took 0.03 seconds to destroy the instance on the hypervisor. [ 582.884613] env[61991]: DEBUG oslo.service.loopingcall [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.884830] env[61991]: DEBUG nova.compute.manager [-] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.884923] env[61991]: DEBUG nova.network.neutron [-] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 582.925766] env[61991]: DEBUG nova.network.neutron [-] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.180188] env[61991]: DEBUG nova.compute.utils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.181655] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 583.181833] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 583.260141] env[61991]: DEBUG nova.policy [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbb0bf057c4c4ee29338848a5ff48b80', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1785e389d40041f3bd72f0c7745dc707', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.364842] env[61991]: INFO nova.scheduler.client.report [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Deleted allocations for instance 8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65 [ 583.436681] env[61991]: DEBUG nova.network.neutron [-] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.623229] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c2ad02-eec2-468d-aae0-443d8ed435dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.630830] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c8aceb-397c-4eba-95ac-85696ffa1f54 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.667221] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62100770-7f78-4ba1-bc05-9925764dec5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.677432] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0574a516-c4d3-44b0-ba53-3b4d7632f562 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.693515] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.699289] env[61991]: DEBUG nova.compute.provider_tree [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.808384] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Successfully created port: a2fd55ab-77f8-436a-9c77-f874f86c95dc {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.879567] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0429872-1cf9-4c44-92c3-77a618336fdf tempest-ImagesOneServerTestJSON-991986383 tempest-ImagesOneServerTestJSON-991986383-project-member] Lock "8c1cdc60-66dc-4e0e-847c-c9a2c23cfd65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.992s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.939823] env[61991]: INFO nova.compute.manager [-] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Took 1.05 seconds to deallocate network for instance. [ 583.942136] env[61991]: DEBUG nova.compute.claims [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.942329] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.176994] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "acef131a-4263-42ad-964e-bb1e0cb21eec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.177421] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "acef131a-4263-42ad-964e-bb1e0cb21eec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.207790] env[61991]: DEBUG nova.scheduler.client.report [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.386668] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.714742] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "fe78a8de-6174-47b3-9271-171c4ba76068" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.715565] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "fe78a8de-6174-47b3-9271-171c4ba76068" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.717160] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.722525] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.722525] env[61991]: ERROR nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Traceback (most recent call last): [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self.driver.spawn(context, instance, image_meta, [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.722525] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] vm_ref = self.build_virtual_machine(instance, [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] for vif in network_info: [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return self._sync_wrapper(fn, *args, **kwargs) [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self.wait() [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self[:] = self._gt.wait() [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return self._exit_event.wait() [ 584.722872] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] result = hub.switch() [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return self.greenlet.switch() [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] result = function(*args, **kwargs) [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] return func(*args, **kwargs) [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] raise e [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] nwinfo = self.network_api.allocate_for_instance( [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.723174] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] created_port_ids = self._update_ports_for_instance( [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] with excutils.save_and_reraise_exception(): [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] self.force_reraise() [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] raise self.value [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] updated_port = self._update_port( [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] _ensure_no_port_binding_failure(port) [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.723497] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] raise exception.PortBindingFailed(port_id=port['id']) [ 584.727349] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] nova.exception.PortBindingFailed: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. [ 584.727349] env[61991]: ERROR nova.compute.manager [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] [ 584.727349] env[61991]: DEBUG nova.compute.utils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.727349] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.050s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.727905] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Build of instance 4ea6d766-cf90-4248-a25f-b3763f1cc7ee was re-scheduled: Binding failed for port 250a68b4-a552-4b12-b6f4-93816196dec8, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.728297] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.728505] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.728643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.728793] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 584.760025] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.760025] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.760025] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.760238] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.760238] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.760238] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.760238] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.760346] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.760515] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.760713] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.760919] env[61991]: DEBUG nova.virt.hardware [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.761864] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c8d2a4-49fb-4603-9f8c-020071d226f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.771061] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f65ea07-3a13-4d65-827c-63bad536f4f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.909954] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.099557] env[61991]: DEBUG nova.compute.manager [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Received event network-changed-a2fd55ab-77f8-436a-9c77-f874f86c95dc {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 585.099907] env[61991]: DEBUG nova.compute.manager [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Refreshing instance network info cache due to event network-changed-a2fd55ab-77f8-436a-9c77-f874f86c95dc. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 585.104480] env[61991]: DEBUG oslo_concurrency.lockutils [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] Acquiring lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.104480] env[61991]: DEBUG oslo_concurrency.lockutils [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] Acquired lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.104480] env[61991]: DEBUG nova.network.neutron [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Refreshing network info cache for port a2fd55ab-77f8-436a-9c77-f874f86c95dc {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 585.257942] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.319945] env[61991]: ERROR nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 585.319945] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.319945] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.319945] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.319945] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.319945] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.319945] env[61991]: ERROR nova.compute.manager raise self.value [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.319945] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.319945] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.319945] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.320390] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.320390] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.320390] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 585.320390] env[61991]: ERROR nova.compute.manager [ 585.320390] env[61991]: Traceback (most recent call last): [ 585.320390] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.320390] env[61991]: listener.cb(fileno) [ 585.320390] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.320390] env[61991]: result = function(*args, **kwargs) [ 585.320390] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.320390] env[61991]: return func(*args, **kwargs) [ 585.320390] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.320390] env[61991]: raise e [ 585.320390] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.320390] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 585.320390] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.320390] env[61991]: created_port_ids = self._update_ports_for_instance( [ 585.320390] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.320390] env[61991]: with excutils.save_and_reraise_exception(): [ 585.320390] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.320390] env[61991]: self.force_reraise() [ 585.320390] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.320390] env[61991]: raise self.value [ 585.320390] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.320390] env[61991]: updated_port = self._update_port( [ 585.320390] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.320390] env[61991]: _ensure_no_port_binding_failure(port) [ 585.320390] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.320390] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.321102] env[61991]: nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 585.321102] env[61991]: Removing descriptor: 17 [ 585.321102] env[61991]: ERROR nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Traceback (most recent call last): [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] yield resources [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self.driver.spawn(context, instance, image_meta, [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.321102] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] vm_ref = self.build_virtual_machine(instance, [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] for vif in network_info: [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return self._sync_wrapper(fn, *args, **kwargs) [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self.wait() [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self[:] = self._gt.wait() [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return self._exit_event.wait() [ 585.321510] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] result = hub.switch() [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return self.greenlet.switch() [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] result = function(*args, **kwargs) [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return func(*args, **kwargs) [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] raise e [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] nwinfo = self.network_api.allocate_for_instance( [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.321895] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] created_port_ids = self._update_ports_for_instance( [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] with excutils.save_and_reraise_exception(): [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self.force_reraise() [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] raise self.value [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] updated_port = self._update_port( [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] _ensure_no_port_binding_failure(port) [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.322250] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] raise exception.PortBindingFailed(port_id=port['id']) [ 585.322509] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 585.322509] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] [ 585.322509] env[61991]: INFO nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Terminating instance [ 585.322509] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquiring lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.344209] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.625159] env[61991]: DEBUG nova.network.neutron [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.672566] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9914a8e4-35ac-4357-ae2f-1cd48ee5a7a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.684036] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b526a0-6161-49b3-84f8-ea2963849b3e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.715256] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfbfdf9-d52a-4196-a1f1-3f44ad51099a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.718676] env[61991]: DEBUG nova.network.neutron [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.725342] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50a8409-7636-42b9-96c0-d07a07798930 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.741142] env[61991]: DEBUG nova.compute.provider_tree [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.846369] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-4ea6d766-cf90-4248-a25f-b3763f1cc7ee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.846416] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.846557] env[61991]: DEBUG nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.846703] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.870200] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.221811] env[61991]: DEBUG oslo_concurrency.lockutils [req-3d8c92ec-c292-4b61-9156-0db288641cab req-cc33f61d-e169-4753-b9cd-e2b84db8f6b7 service nova] Releasing lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.222234] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquired lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.222414] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.244151] env[61991]: DEBUG nova.scheduler.client.report [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.374143] env[61991]: DEBUG nova.network.neutron [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.749257] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.758762] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.759401] env[61991]: ERROR nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Traceback (most recent call last): [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self.driver.spawn(context, instance, image_meta, [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] vm_ref = self.build_virtual_machine(instance, [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.759401] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] for vif in network_info: [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return self._sync_wrapper(fn, *args, **kwargs) [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self.wait() [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self[:] = self._gt.wait() [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return self._exit_event.wait() [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] result = hub.switch() [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.759857] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return self.greenlet.switch() [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] result = function(*args, **kwargs) [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] return func(*args, **kwargs) [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] raise e [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] nwinfo = self.network_api.allocate_for_instance( [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] created_port_ids = self._update_ports_for_instance( [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] with excutils.save_and_reraise_exception(): [ 586.760202] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] self.force_reraise() [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] raise self.value [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] updated_port = self._update_port( [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] _ensure_no_port_binding_failure(port) [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] raise exception.PortBindingFailed(port_id=port['id']) [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] nova.exception.PortBindingFailed: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. [ 586.760503] env[61991]: ERROR nova.compute.manager [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] [ 586.760851] env[61991]: DEBUG nova.compute.utils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.761558] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.004s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.763813] env[61991]: INFO nova.compute.claims [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.770028] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Build of instance 9043ea71-0207-42c7-8907-0ae61f210fee was re-scheduled: Binding failed for port 2bd1b153-8242-407a-9733-d5565cb6c677, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.770028] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.770028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquiring lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.770028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Acquired lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.770178] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.860452] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.877261] env[61991]: INFO nova.compute.manager [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: 4ea6d766-cf90-4248-a25f-b3763f1cc7ee] Took 1.03 seconds to deallocate network for instance. [ 587.134085] env[61991]: DEBUG nova.compute.manager [req-e1b27d33-8ed1-4db2-bec2-57c6ac89db13 req-e996939f-58d1-4efd-b9aa-4e89bfb51cf9 service nova] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Received event network-vif-deleted-a2fd55ab-77f8-436a-9c77-f874f86c95dc {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.247766] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.247995] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.287180] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.363773] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Releasing lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.364047] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.364278] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 587.364570] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4171a9f2-5eec-4a15-b43c-cc04749b645f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.372318] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.376197] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5317835-266c-4dd7-b3f2-c21432f006a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.387882] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Releasing lock "refresh_cache-9043ea71-0207-42c7-8907-0ae61f210fee" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.388902] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.388902] env[61991]: DEBUG nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.388902] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.406894] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39ba999e-8740-4f9a-8b08-516e5d80d233 could not be found. [ 587.407181] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.407823] env[61991]: INFO nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.407823] env[61991]: DEBUG oslo.service.loopingcall [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.408390] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.410069] env[61991]: DEBUG nova.compute.manager [-] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.410069] env[61991]: DEBUG nova.network.neutron [-] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.424219] env[61991]: DEBUG nova.network.neutron [-] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.911819] env[61991]: INFO nova.scheduler.client.report [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Deleted allocations for instance 4ea6d766-cf90-4248-a25f-b3763f1cc7ee [ 587.917226] env[61991]: DEBUG nova.network.neutron [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.926782] env[61991]: DEBUG nova.network.neutron [-] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.142979] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f888a9e-e834-487e-a8ff-b9e336a880e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.150463] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057f669f-5385-43fa-878a-53c35653cf0b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.179956] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd9bfbb-a202-46c1-81e2-fd2c324758cf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.187302] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a276b47a-37e4-4d4b-9fa7-358a0dd07f63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.200842] env[61991]: DEBUG nova.compute.provider_tree [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.423160] env[61991]: INFO nova.compute.manager [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] [instance: 9043ea71-0207-42c7-8907-0ae61f210fee] Took 1.03 seconds to deallocate network for instance. [ 588.427241] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a82db7ea-d040-4bcb-9bd5-23304d035718 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "4ea6d766-cf90-4248-a25f-b3763f1cc7ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.229s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.430300] env[61991]: INFO nova.compute.manager [-] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Took 1.02 seconds to deallocate network for instance. [ 588.433169] env[61991]: DEBUG nova.compute.claims [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.433336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.703487] env[61991]: DEBUG nova.scheduler.client.report [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.932085] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.211891] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.211891] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.213047] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.483s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.459153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.464266] env[61991]: INFO nova.scheduler.client.report [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Deleted allocations for instance 9043ea71-0207-42c7-8907-0ae61f210fee [ 589.721720] env[61991]: DEBUG nova.compute.utils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.726018] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 589.726827] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 589.797282] env[61991]: DEBUG nova.policy [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d1e7573f5ca46f98f4c1ecc4ec0db82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1adf38ed214445c9a9ac7307338e676d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.973374] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d709ce30-5330-4ade-9c41-9fd11fc18b17 tempest-FloatingIPsAssociationTestJSON-185525490 tempest-FloatingIPsAssociationTestJSON-185525490-project-member] Lock "9043ea71-0207-42c7-8907-0ae61f210fee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.251s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.094757] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7292266c-96ce-4c9d-a57a-962ea598ca86 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.102473] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b4014d-ad41-4b8e-9b21-4bbddbac9b5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.138408] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3989d2-1963-44e8-8fc8-d5aed2c34557 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.148126] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43c71e9-5da4-41ff-8e8f-e4d6dcf0f594 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.166206] env[61991]: DEBUG nova.compute.provider_tree [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.228145] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.240599] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "c14d99a3-950d-44d0-b330-3031a1f8a2be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.240599] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.252053] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Successfully created port: 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.482217] env[61991]: DEBUG nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.669086] env[61991]: DEBUG nova.scheduler.client.report [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.781053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "527561af-ba19-447f-a4ee-618824b9ce0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.781053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "527561af-ba19-447f-a4ee-618824b9ce0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.017476] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.159600] env[61991]: DEBUG nova.compute.manager [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Received event network-changed-2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.160693] env[61991]: DEBUG nova.compute.manager [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Refreshing instance network info cache due to event network-changed-2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.160807] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] Acquiring lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.160908] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] Acquired lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.161084] env[61991]: DEBUG nova.network.neutron [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Refreshing network info cache for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 591.176019] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.961s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.176019] env[61991]: ERROR nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Traceback (most recent call last): [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self.driver.spawn(context, instance, image_meta, [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.176019] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] vm_ref = self.build_virtual_machine(instance, [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] for vif in network_info: [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return self._sync_wrapper(fn, *args, **kwargs) [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self.wait() [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self[:] = self._gt.wait() [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return self._exit_event.wait() [ 591.177084] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] result = hub.switch() [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return self.greenlet.switch() [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] result = function(*args, **kwargs) [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] return func(*args, **kwargs) [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] raise e [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] nwinfo = self.network_api.allocate_for_instance( [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.177399] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] created_port_ids = self._update_ports_for_instance( [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] with excutils.save_and_reraise_exception(): [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] self.force_reraise() [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] raise self.value [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] updated_port = self._update_port( [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] _ensure_no_port_binding_failure(port) [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.177678] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] raise exception.PortBindingFailed(port_id=port['id']) [ 591.177954] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] nova.exception.PortBindingFailed: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. [ 591.177954] env[61991]: ERROR nova.compute.manager [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] [ 591.177954] env[61991]: DEBUG nova.compute.utils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.179331] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.205s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.183374] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Build of instance 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300 was re-scheduled: Binding failed for port 121bf0e8-89de-48b1-8d67-4d960901fbf2, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.184203] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.184472] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquiring lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.184712] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Acquired lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.185154] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.238481] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.271589] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.271922] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.272146] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.272428] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.272480] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.272623] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.272821] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.273056] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.273311] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.274040] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.274277] env[61991]: DEBUG nova.virt.hardware [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.275182] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334ee0d3-a861-4268-a042-a4194f2adb21 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.283247] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cff5095-05fa-42d4-9d9a-6a7a2c146483 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.380985] env[61991]: ERROR nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 591.380985] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.380985] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.380985] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.380985] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.380985] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.380985] env[61991]: ERROR nova.compute.manager raise self.value [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.380985] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.380985] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.380985] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.381434] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.381434] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.381434] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 591.381434] env[61991]: ERROR nova.compute.manager [ 591.381434] env[61991]: Traceback (most recent call last): [ 591.381434] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.381434] env[61991]: listener.cb(fileno) [ 591.381434] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.381434] env[61991]: result = function(*args, **kwargs) [ 591.381434] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.381434] env[61991]: return func(*args, **kwargs) [ 591.381434] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.381434] env[61991]: raise e [ 591.381434] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.381434] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 591.381434] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.381434] env[61991]: created_port_ids = self._update_ports_for_instance( [ 591.381434] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.381434] env[61991]: with excutils.save_and_reraise_exception(): [ 591.381434] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.381434] env[61991]: self.force_reraise() [ 591.381434] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.381434] env[61991]: raise self.value [ 591.381434] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.381434] env[61991]: updated_port = self._update_port( [ 591.381434] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.381434] env[61991]: _ensure_no_port_binding_failure(port) [ 591.381434] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.381434] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.382199] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 591.382199] env[61991]: Removing descriptor: 18 [ 591.382199] env[61991]: ERROR nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Traceback (most recent call last): [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] yield resources [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self.driver.spawn(context, instance, image_meta, [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.382199] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] vm_ref = self.build_virtual_machine(instance, [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] for vif in network_info: [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return self._sync_wrapper(fn, *args, **kwargs) [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self.wait() [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self[:] = self._gt.wait() [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return self._exit_event.wait() [ 591.382573] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] result = hub.switch() [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return self.greenlet.switch() [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] result = function(*args, **kwargs) [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return func(*args, **kwargs) [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] raise e [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] nwinfo = self.network_api.allocate_for_instance( [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.382942] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] created_port_ids = self._update_ports_for_instance( [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] with excutils.save_and_reraise_exception(): [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self.force_reraise() [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] raise self.value [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] updated_port = self._update_port( [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] _ensure_no_port_binding_failure(port) [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.383295] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] raise exception.PortBindingFailed(port_id=port['id']) [ 591.383613] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 591.383613] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] [ 591.383613] env[61991]: INFO nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Terminating instance [ 591.384474] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquiring lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.680060] env[61991]: DEBUG nova.network.neutron [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.729779] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.858532] env[61991]: DEBUG nova.network.neutron [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.874769] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.113264] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a89252-f74e-4f9c-933c-d69172d5af43 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.120969] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d6f037-6ad8-4236-a319-301b301005b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.149712] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e620b77b-e2f6-4883-bcd1-03874c171e17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.157023] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb872f7-3972-4dab-a890-3909c3d22bc3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.170593] env[61991]: DEBUG nova.compute.provider_tree [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.361651] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6953f70-a09d-45e0-bc37-3cef47123f7b req-6a5f2258-161d-4643-8505-5c564376aa86 service nova] Releasing lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.362084] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquired lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.362269] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.377088] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Releasing lock "refresh_cache-2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.377340] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.377547] env[61991]: DEBUG nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.377712] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.394684] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.673873] env[61991]: DEBUG nova.scheduler.client.report [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.887831] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.897640] env[61991]: DEBUG nova.network.neutron [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.996992] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.179825] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.181192] env[61991]: ERROR nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Traceback (most recent call last): [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self.driver.spawn(context, instance, image_meta, [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] vm_ref = self.build_virtual_machine(instance, [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.181192] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] for vif in network_info: [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return self._sync_wrapper(fn, *args, **kwargs) [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self.wait() [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self[:] = self._gt.wait() [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return self._exit_event.wait() [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] result = hub.switch() [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.181494] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return self.greenlet.switch() [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] result = function(*args, **kwargs) [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] return func(*args, **kwargs) [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] raise e [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] nwinfo = self.network_api.allocate_for_instance( [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] created_port_ids = self._update_ports_for_instance( [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] with excutils.save_and_reraise_exception(): [ 593.182066] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] self.force_reraise() [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] raise self.value [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] updated_port = self._update_port( [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] _ensure_no_port_binding_failure(port) [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] raise exception.PortBindingFailed(port_id=port['id']) [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] nova.exception.PortBindingFailed: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. [ 593.182359] env[61991]: ERROR nova.compute.manager [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] [ 593.182606] env[61991]: DEBUG nova.compute.utils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.182606] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.799s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.185814] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Build of instance 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea was re-scheduled: Binding failed for port 8db2d35c-9b14-4476-8e29-b5a83c9b70e0, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.186260] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.186486] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.187579] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.187579] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.193941] env[61991]: DEBUG nova.compute.manager [req-378e123f-f7ef-420e-9000-29b4429b071d req-239a8adf-6e34-4b98-8b93-2cf85d885f20 service nova] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Received event network-vif-deleted-2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.399683] env[61991]: INFO nova.compute.manager [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] [instance: 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300] Took 1.02 seconds to deallocate network for instance. [ 593.500200] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Releasing lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.503017] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.503017] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.503017] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-232334fb-36fc-45a2-baf1-507865b56856 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.514336] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679c5293-cc61-4456-be44-ee6a8a95326c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.537111] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69db1562-7169-4ebc-899d-c54ee6e03384 could not be found. [ 593.537347] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.537527] env[61991]: INFO nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.537770] env[61991]: DEBUG oslo.service.loopingcall [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.537982] env[61991]: DEBUG nova.compute.manager [-] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.538095] env[61991]: DEBUG nova.network.neutron [-] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.552850] env[61991]: DEBUG nova.network.neutron [-] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.707904] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.792961] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.055044] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c0fc40-430b-43c9-bba4-97c355950c29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.059142] env[61991]: DEBUG nova.network.neutron [-] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.066860] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcae711-780a-478e-8d28-25a53b0573e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.097912] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659004e5-af9b-4744-a0d0-ad2b330e7f2d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.105072] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be41d708-6b1d-48a2-8943-85bb2317fde8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.118358] env[61991]: DEBUG nova.compute.provider_tree [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.295436] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.295688] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.295974] env[61991]: DEBUG nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.296057] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.328052] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.434940] env[61991]: INFO nova.scheduler.client.report [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Deleted allocations for instance 2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300 [ 594.566071] env[61991]: INFO nova.compute.manager [-] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Took 1.02 seconds to deallocate network for instance. [ 594.566071] env[61991]: DEBUG nova.compute.claims [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.566414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.622617] env[61991]: DEBUG nova.scheduler.client.report [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.830819] env[61991]: DEBUG nova.network.neutron [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.945025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8fa05f15-f1da-4bf3-9ec5-b0491060fb8a tempest-AttachInterfacesV270Test-1833794252 tempest-AttachInterfacesV270Test-1833794252-project-member] Lock "2bb2ff5e-eccc-4c6b-8058-3b8c6eec8300" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.153s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.126364] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.944s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.126999] env[61991]: ERROR nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Traceback (most recent call last): [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self.driver.spawn(context, instance, image_meta, [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] vm_ref = self.build_virtual_machine(instance, [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.126999] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] for vif in network_info: [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return self._sync_wrapper(fn, *args, **kwargs) [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self.wait() [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self[:] = self._gt.wait() [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return self._exit_event.wait() [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] result = hub.switch() [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.127284] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return self.greenlet.switch() [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] result = function(*args, **kwargs) [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] return func(*args, **kwargs) [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] raise e [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] nwinfo = self.network_api.allocate_for_instance( [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] created_port_ids = self._update_ports_for_instance( [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] with excutils.save_and_reraise_exception(): [ 595.127565] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] self.force_reraise() [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] raise self.value [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] updated_port = self._update_port( [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] _ensure_no_port_binding_failure(port) [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] raise exception.PortBindingFailed(port_id=port['id']) [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] nova.exception.PortBindingFailed: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. [ 595.127841] env[61991]: ERROR nova.compute.manager [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] [ 595.128109] env[61991]: DEBUG nova.compute.utils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.128951] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.596s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.130485] env[61991]: INFO nova.compute.claims [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.133254] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Build of instance 3a74ce10-43cb-4c56-a47a-941c8effa14a was re-scheduled: Binding failed for port 7a72e257-b290-4f08-965e-4491625c0376, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 595.133711] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 595.133933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.134126] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquired lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.134293] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.334401] env[61991]: INFO nova.compute.manager [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea] Took 1.04 seconds to deallocate network for instance. [ 595.446493] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.664721] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.800336] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.972519] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.219032] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "43cb61e7-1748-40d9-a287-1179c8219c2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.219032] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.302342] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Releasing lock "refresh_cache-3a74ce10-43cb-4c56-a47a-941c8effa14a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.302691] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.302753] env[61991]: DEBUG nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.302928] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.325406] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.369948] env[61991]: INFO nova.scheduler.client.report [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted allocations for instance 8f8a933b-ce41-4c22-aafd-8e730f4fb1ea [ 596.589843] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef95ddb-862a-4581-815e-04748bab7b4e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.597624] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3f9927-f917-4fda-ba40-eddb835be596 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.629749] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e413d26-c24e-41fe-8018-6e17056f4699 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.637350] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761566f4-abc5-4ab1-8b9f-dd34314e25a6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.651162] env[61991]: DEBUG nova.compute.provider_tree [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.830483] env[61991]: DEBUG nova.network.neutron [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.877943] env[61991]: DEBUG oslo_concurrency.lockutils [None req-acb1ef89-ebf7-43a2-8194-8cef20b86908 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "8f8a933b-ce41-4c22-aafd-8e730f4fb1ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.123s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.155117] env[61991]: DEBUG nova.scheduler.client.report [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.333452] env[61991]: INFO nova.compute.manager [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 3a74ce10-43cb-4c56-a47a-941c8effa14a] Took 1.03 seconds to deallocate network for instance. [ 597.380184] env[61991]: DEBUG nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.659916] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.660565] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.663231] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.721s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.905440] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.049236] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "1715711a-72c5-4e86-88e0-cecf922cb42f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.049452] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.167386] env[61991]: DEBUG nova.compute.utils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.171727] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.171896] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 598.226164] env[61991]: DEBUG nova.policy [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.373080] env[61991]: INFO nova.scheduler.client.report [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Deleted allocations for instance 3a74ce10-43cb-4c56-a47a-941c8effa14a [ 598.540926] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Successfully created port: 0e78c05b-7e20-49ef-8118-cd1600aa4d23 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.548383] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e16478-36c7-4eb5-a17f-af506bfa8bc7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.555135] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc2d3d7-9975-44b2-ac67-9f69747db2d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.588288] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855516bd-15fc-4c8a-9abb-e6a2442a182f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.596705] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc8a80a-5e58-4a3f-b38e-c93dd06c54c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.612969] env[61991]: DEBUG nova.compute.provider_tree [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.672726] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.882976] env[61991]: DEBUG oslo_concurrency.lockutils [None req-426633f3-8aba-4808-8d2d-0db32b4bf6c6 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "3a74ce10-43cb-4c56-a47a-941c8effa14a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.146s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.119828] env[61991]: DEBUG nova.scheduler.client.report [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.389244] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.488556] env[61991]: DEBUG nova.compute.manager [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] [instance: b7237462-348e-4e88-b415-a301499fab83] Received event network-changed-0e78c05b-7e20-49ef-8118-cd1600aa4d23 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.488556] env[61991]: DEBUG nova.compute.manager [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] [instance: b7237462-348e-4e88-b415-a301499fab83] Refreshing instance network info cache due to event network-changed-0e78c05b-7e20-49ef-8118-cd1600aa4d23. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 599.488680] env[61991]: DEBUG oslo_concurrency.lockutils [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] Acquiring lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.488816] env[61991]: DEBUG oslo_concurrency.lockutils [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] Acquired lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.489983] env[61991]: DEBUG nova.network.neutron [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] [instance: b7237462-348e-4e88-b415-a301499fab83] Refreshing network info cache for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 599.626061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.626855] env[61991]: ERROR nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Traceback (most recent call last): [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self.driver.spawn(context, instance, image_meta, [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] vm_ref = self.build_virtual_machine(instance, [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.626855] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] for vif in network_info: [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return self._sync_wrapper(fn, *args, **kwargs) [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self.wait() [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self[:] = self._gt.wait() [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return self._exit_event.wait() [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] result = hub.switch() [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.627190] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return self.greenlet.switch() [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] result = function(*args, **kwargs) [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] return func(*args, **kwargs) [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] raise e [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] nwinfo = self.network_api.allocate_for_instance( [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] created_port_ids = self._update_ports_for_instance( [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] with excutils.save_and_reraise_exception(): [ 599.627545] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] self.force_reraise() [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] raise self.value [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] updated_port = self._update_port( [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] _ensure_no_port_binding_failure(port) [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] raise exception.PortBindingFailed(port_id=port['id']) [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] nova.exception.PortBindingFailed: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. [ 599.627884] env[61991]: ERROR nova.compute.manager [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] [ 599.632124] env[61991]: DEBUG nova.compute.utils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.632124] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.720s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.632124] env[61991]: INFO nova.compute.claims [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.634396] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Build of instance 4ce0f385-cee1-4afd-ac38-5624623953ac was re-scheduled: Binding failed for port 99edbbbc-52da-476d-8302-b494ffa0c42d, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.634843] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.635082] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.635281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.635699] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.686168] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.694366] env[61991]: ERROR nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 599.694366] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.694366] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.694366] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.694366] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.694366] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.694366] env[61991]: ERROR nova.compute.manager raise self.value [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.694366] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.694366] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.694366] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.694785] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.694785] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.694785] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 599.694785] env[61991]: ERROR nova.compute.manager [ 599.694785] env[61991]: Traceback (most recent call last): [ 599.694785] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.694785] env[61991]: listener.cb(fileno) [ 599.694785] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.694785] env[61991]: result = function(*args, **kwargs) [ 599.694785] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.694785] env[61991]: return func(*args, **kwargs) [ 599.694785] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.694785] env[61991]: raise e [ 599.694785] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.694785] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 599.694785] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.694785] env[61991]: created_port_ids = self._update_ports_for_instance( [ 599.694785] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.694785] env[61991]: with excutils.save_and_reraise_exception(): [ 599.694785] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.694785] env[61991]: self.force_reraise() [ 599.694785] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.694785] env[61991]: raise self.value [ 599.694785] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.694785] env[61991]: updated_port = self._update_port( [ 599.694785] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.694785] env[61991]: _ensure_no_port_binding_failure(port) [ 599.694785] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.694785] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.695438] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 599.695438] env[61991]: Removing descriptor: 17 [ 599.720544] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.721354] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.721566] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.721762] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.721905] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.722066] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.722271] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.722423] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.722581] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.722735] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.722975] env[61991]: DEBUG nova.virt.hardware [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.724156] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c1c0b5-af03-4ad3-abae-06ad12d50278 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.733752] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527899fd-3f95-4161-a467-6a2fd8692c68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.747879] env[61991]: ERROR nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] Traceback (most recent call last): [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] yield resources [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self.driver.spawn(context, instance, image_meta, [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] vm_ref = self.build_virtual_machine(instance, [ 599.747879] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] for vif in network_info: [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] return self._sync_wrapper(fn, *args, **kwargs) [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self.wait() [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self[:] = self._gt.wait() [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] return self._exit_event.wait() [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 599.748227] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] current.throw(*self._exc) [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] result = function(*args, **kwargs) [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] return func(*args, **kwargs) [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] raise e [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] nwinfo = self.network_api.allocate_for_instance( [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] created_port_ids = self._update_ports_for_instance( [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] with excutils.save_and_reraise_exception(): [ 599.748544] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self.force_reraise() [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] raise self.value [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] updated_port = self._update_port( [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] _ensure_no_port_binding_failure(port) [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] raise exception.PortBindingFailed(port_id=port['id']) [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 599.748858] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] [ 599.748858] env[61991]: INFO nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Terminating instance [ 599.750578] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.919397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.027904] env[61991]: DEBUG nova.network.neutron [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.165583] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.200101] env[61991]: DEBUG nova.network.neutron [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] [instance: b7237462-348e-4e88-b415-a301499fab83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.333152] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.704994] env[61991]: DEBUG oslo_concurrency.lockutils [req-88625423-525a-4abe-ac0a-d2880d8d2184 req-288269cf-24b2-4bf8-92c5-9a869c8d3588 service nova] Releasing lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.707829] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.708631] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.840693] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "refresh_cache-4ce0f385-cee1-4afd-ac38-5624623953ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.841339] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.841339] env[61991]: DEBUG nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.841499] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.876416] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.068805] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665b4902-1ab3-4c2b-bf3c-645e45b83504 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.077487] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891217ed-95a0-4840-87ca-333877acfe92 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.735087] env[61991]: DEBUG nova.network.neutron [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.739445] env[61991]: DEBUG nova.compute.manager [req-72a96dd8-81cb-4129-b7cb-4a4002238cf5 req-f33689bc-c72c-4ba2-9bdd-9ab434866218 service nova] [instance: b7237462-348e-4e88-b415-a301499fab83] Received event network-vif-deleted-0e78c05b-7e20-49ef-8118-cd1600aa4d23 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.739445] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e340a2-8642-4933-a021-23cb72441e45 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.748291] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b812b3-eb62-4774-8f31-def28c5111f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.754042] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.763487] env[61991]: DEBUG nova.compute.provider_tree [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.859524] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.930935] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.931974] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.238959] env[61991]: INFO nova.compute.manager [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: 4ce0f385-cee1-4afd-ac38-5624623953ac] Took 1.40 seconds to deallocate network for instance. [ 602.266125] env[61991]: DEBUG nova.scheduler.client.report [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.362672] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.363248] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.363445] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.363734] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cae149aa-5c0f-4593-90b9-61816e7ee296 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.372054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3fa908-5c69-4d25-8403-7b0dae10aa0a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.392415] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b7237462-348e-4e88-b415-a301499fab83 could not be found. [ 602.392665] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.392861] env[61991]: INFO nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Took 0.03 seconds to destroy the instance on the hypervisor. [ 602.393498] env[61991]: DEBUG oslo.service.loopingcall [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.393727] env[61991]: DEBUG nova.compute.manager [-] [instance: b7237462-348e-4e88-b415-a301499fab83] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.393823] env[61991]: DEBUG nova.network.neutron [-] [instance: b7237462-348e-4e88-b415-a301499fab83] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.408738] env[61991]: DEBUG nova.network.neutron [-] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.467372] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.467603] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.770944] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.141s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.771963] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.774396] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.341s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.911757] env[61991]: DEBUG nova.network.neutron [-] [instance: b7237462-348e-4e88-b415-a301499fab83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.266200] env[61991]: INFO nova.scheduler.client.report [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleted allocations for instance 4ce0f385-cee1-4afd-ac38-5624623953ac [ 603.279526] env[61991]: DEBUG nova.compute.utils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.281217] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.281490] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 603.338814] env[61991]: DEBUG nova.policy [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e6810d3aced4a44bbfe9ba49270bc35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a18b2a73f7643e1b89c9660a770e9da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.417076] env[61991]: INFO nova.compute.manager [-] [instance: b7237462-348e-4e88-b415-a301499fab83] Took 1.02 seconds to deallocate network for instance. [ 603.420564] env[61991]: DEBUG nova.compute.claims [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.420751] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.620859] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8361125-98e2-4fa5-8f95-beab99d22fdf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.628684] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Successfully created port: f1e8449e-af0e-44f4-a850-8f4efe4b1e40 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.632048] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48079fb8-0206-4eea-a3ea-fe3f7beec4f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.662024] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53eadda-9997-4479-8e11-3b2a69323122 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.668792] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fe0e01-7552-4050-b1e6-db7ce5fcb392 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.682492] env[61991]: DEBUG nova.compute.provider_tree [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.774738] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3ae7959e-1474-4022-a3a6-05dfaf2dd2b9 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "4ce0f385-cee1-4afd-ac38-5624623953ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.868s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.787904] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.186222] env[61991]: DEBUG nova.scheduler.client.report [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.279441] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.554308] env[61991]: DEBUG nova.compute.manager [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Received event network-changed-f1e8449e-af0e-44f4-a850-8f4efe4b1e40 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.554308] env[61991]: DEBUG nova.compute.manager [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Refreshing instance network info cache due to event network-changed-f1e8449e-af0e-44f4-a850-8f4efe4b1e40. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.555126] env[61991]: DEBUG oslo_concurrency.lockutils [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] Acquiring lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.555126] env[61991]: DEBUG oslo_concurrency.lockutils [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] Acquired lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.555126] env[61991]: DEBUG nova.network.neutron [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Refreshing network info cache for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.699258] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.699258] env[61991]: ERROR nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Traceback (most recent call last): [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self.driver.spawn(context, instance, image_meta, [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.699258] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] vm_ref = self.build_virtual_machine(instance, [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] for vif in network_info: [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return self._sync_wrapper(fn, *args, **kwargs) [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self.wait() [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self[:] = self._gt.wait() [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return self._exit_event.wait() [ 604.699563] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] result = hub.switch() [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return self.greenlet.switch() [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] result = function(*args, **kwargs) [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] return func(*args, **kwargs) [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] raise e [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] nwinfo = self.network_api.allocate_for_instance( [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.699918] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] created_port_ids = self._update_ports_for_instance( [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] with excutils.save_and_reraise_exception(): [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] self.force_reraise() [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] raise self.value [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] updated_port = self._update_port( [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] _ensure_no_port_binding_failure(port) [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.700284] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] raise exception.PortBindingFailed(port_id=port['id']) [ 604.700651] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] nova.exception.PortBindingFailed: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. [ 604.700651] env[61991]: ERROR nova.compute.manager [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] [ 604.700651] env[61991]: DEBUG nova.compute.utils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.700651] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.240s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.701669] env[61991]: INFO nova.compute.claims [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.704622] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Build of instance 39ba999e-8740-4f9a-8b08-516e5d80d233 was re-scheduled: Binding failed for port a2fd55ab-77f8-436a-9c77-f874f86c95dc, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.704763] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.704937] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquiring lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.705515] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Acquired lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.705515] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.735108] env[61991]: ERROR nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 604.735108] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.735108] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.735108] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.735108] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.735108] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.735108] env[61991]: ERROR nova.compute.manager raise self.value [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.735108] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 604.735108] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.735108] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 604.735592] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.735592] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 604.735592] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 604.735592] env[61991]: ERROR nova.compute.manager [ 604.735592] env[61991]: Traceback (most recent call last): [ 604.735592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 604.735592] env[61991]: listener.cb(fileno) [ 604.735592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.735592] env[61991]: result = function(*args, **kwargs) [ 604.735592] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.735592] env[61991]: return func(*args, **kwargs) [ 604.735592] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.735592] env[61991]: raise e [ 604.735592] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.735592] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 604.735592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.735592] env[61991]: created_port_ids = self._update_ports_for_instance( [ 604.735592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.735592] env[61991]: with excutils.save_and_reraise_exception(): [ 604.735592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.735592] env[61991]: self.force_reraise() [ 604.735592] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.735592] env[61991]: raise self.value [ 604.735592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.735592] env[61991]: updated_port = self._update_port( [ 604.735592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.735592] env[61991]: _ensure_no_port_binding_failure(port) [ 604.735592] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.735592] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 604.736244] env[61991]: nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 604.736244] env[61991]: Removing descriptor: 18 [ 604.804456] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.808521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.830055] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.830227] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.830426] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.830648] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.831188] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.831188] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.831456] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.831692] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.831900] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.832120] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.832341] env[61991]: DEBUG nova.virt.hardware [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.833264] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13aa4153-0693-4ad9-ab64-506560a37dec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.841650] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5a653d-c7dc-46db-853e-24ec9025160e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.855637] env[61991]: ERROR nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Traceback (most recent call last): [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] yield resources [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self.driver.spawn(context, instance, image_meta, [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] vm_ref = self.build_virtual_machine(instance, [ 604.855637] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] for vif in network_info: [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] return self._sync_wrapper(fn, *args, **kwargs) [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self.wait() [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self[:] = self._gt.wait() [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] return self._exit_event.wait() [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 604.855989] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] current.throw(*self._exc) [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] result = function(*args, **kwargs) [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] return func(*args, **kwargs) [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] raise e [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] nwinfo = self.network_api.allocate_for_instance( [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] created_port_ids = self._update_ports_for_instance( [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] with excutils.save_and_reraise_exception(): [ 604.856335] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self.force_reraise() [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] raise self.value [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] updated_port = self._update_port( [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] _ensure_no_port_binding_failure(port) [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] raise exception.PortBindingFailed(port_id=port['id']) [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 604.856647] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] [ 604.856647] env[61991]: INFO nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Terminating instance [ 604.858532] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.081746] env[61991]: DEBUG nova.network.neutron [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.211752] env[61991]: DEBUG nova.network.neutron [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.236235] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.387308] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.702037] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 605.702434] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 605.718241] env[61991]: DEBUG oslo_concurrency.lockutils [req-f37f8e23-543d-41a7-85fc-9bd0070b3220 req-ae7a9754-103e-43a1-9647-3ca96cf8b6ac service nova] Releasing lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.720213] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.720213] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.890530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Releasing lock "refresh_cache-39ba999e-8740-4f9a-8b08-516e5d80d233" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.890530] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.890530] env[61991]: DEBUG nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.890530] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.969866] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.089991] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab67cf51-12f6-4dee-9524-1b71a971a517 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.098147] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9121483-4e40-4766-8f7d-c64c9f4563e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.127071] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c77a025d-67c9-4ee3-a73f-a9ae965b8412 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.134376] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45ecff2-6594-4cbd-b760-056ae8609b5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.147131] env[61991]: DEBUG nova.compute.provider_tree [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.207523] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.207692] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 606.207808] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 606.242354] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.323252] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.472220] env[61991]: DEBUG nova.network.neutron [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.581783] env[61991]: DEBUG nova.compute.manager [req-0f07860f-5cd8-4a08-ad89-6fb2436c661f req-7d4afcf1-c5df-4b75-a48f-665100af105a service nova] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Received event network-vif-deleted-f1e8449e-af0e-44f4-a850-8f4efe4b1e40 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.650769] env[61991]: DEBUG nova.scheduler.client.report [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.712194] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 606.712194] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: b7237462-348e-4e88-b415-a301499fab83] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 606.712194] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 606.712357] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 606.712395] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Didn't find any instances for network info cache update. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 606.713345] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.713419] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.714058] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.714058] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.714058] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.714058] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.714058] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 606.714271] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 606.789209] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.789432] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.827359] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.827768] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.827977] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.828287] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5ccee89-e023-4394-a642-4809ebab4148 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.837252] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a404e08-b753-4ae7-afb0-3a7c9ead4623 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.858443] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 22be69a5-6405-4e8c-8cdd-0843847f9228 could not be found. [ 606.858656] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.858833] env[61991]: INFO nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Took 0.03 seconds to destroy the instance on the hypervisor. [ 606.859086] env[61991]: DEBUG oslo.service.loopingcall [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.859349] env[61991]: DEBUG nova.compute.manager [-] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.859444] env[61991]: DEBUG nova.network.neutron [-] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.873691] env[61991]: DEBUG nova.network.neutron [-] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.974545] env[61991]: INFO nova.compute.manager [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] [instance: 39ba999e-8740-4f9a-8b08-516e5d80d233] Took 1.08 seconds to deallocate network for instance. [ 607.155027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.155503] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.158302] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.141s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.159657] env[61991]: INFO nova.compute.claims [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.217054] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.376529] env[61991]: DEBUG nova.network.neutron [-] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.663817] env[61991]: DEBUG nova.compute.utils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.667251] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.667251] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.737200] env[61991]: DEBUG nova.policy [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3def0d3763441d8a0f1b248f9b29c14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '12da0c1e6ea741a5b4e84c58a9b09ce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.881799] env[61991]: INFO nova.compute.manager [-] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Took 1.02 seconds to deallocate network for instance. [ 607.884293] env[61991]: DEBUG nova.compute.claims [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.884468] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.016174] env[61991]: INFO nova.scheduler.client.report [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Deleted allocations for instance 39ba999e-8740-4f9a-8b08-516e5d80d233 [ 608.022876] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Successfully created port: 702476ab-f514-40a7-ad38-44acef030059 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.179368] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.527988] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91bf2342-72eb-4fad-9c01-d3aa403bf220 tempest-ServersTestManualDisk-760144761 tempest-ServersTestManualDisk-760144761-project-member] Lock "39ba999e-8740-4f9a-8b08-516e5d80d233" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.329s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.536374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cd398f-7bf1-4200-9bba-4cfa5784378a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.544907] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a3a1de-639d-4dfe-8bbf-813420d74ab3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.581240] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6622db99-6488-4b14-a5bb-71c1c3216704 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.592480] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdf8562-206c-4fb9-925f-66f3bf1cfeec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.607942] env[61991]: DEBUG nova.compute.provider_tree [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.908101] env[61991]: DEBUG nova.compute.manager [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Received event network-changed-702476ab-f514-40a7-ad38-44acef030059 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.908101] env[61991]: DEBUG nova.compute.manager [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Refreshing instance network info cache due to event network-changed-702476ab-f514-40a7-ad38-44acef030059. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 608.908238] env[61991]: DEBUG oslo_concurrency.lockutils [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] Acquiring lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.908323] env[61991]: DEBUG oslo_concurrency.lockutils [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] Acquired lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.908489] env[61991]: DEBUG nova.network.neutron [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Refreshing network info cache for port 702476ab-f514-40a7-ad38-44acef030059 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 609.032421] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.048481] env[61991]: ERROR nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 609.048481] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.048481] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.048481] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.048481] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.048481] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.048481] env[61991]: ERROR nova.compute.manager raise self.value [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.048481] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.048481] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.048481] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.049505] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.049505] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.049505] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 609.049505] env[61991]: ERROR nova.compute.manager [ 609.049505] env[61991]: Traceback (most recent call last): [ 609.049505] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.049505] env[61991]: listener.cb(fileno) [ 609.049505] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.049505] env[61991]: result = function(*args, **kwargs) [ 609.049505] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.049505] env[61991]: return func(*args, **kwargs) [ 609.049505] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.049505] env[61991]: raise e [ 609.049505] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.049505] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 609.049505] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.049505] env[61991]: created_port_ids = self._update_ports_for_instance( [ 609.049505] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.049505] env[61991]: with excutils.save_and_reraise_exception(): [ 609.049505] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.049505] env[61991]: self.force_reraise() [ 609.049505] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.049505] env[61991]: raise self.value [ 609.049505] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.049505] env[61991]: updated_port = self._update_port( [ 609.049505] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.049505] env[61991]: _ensure_no_port_binding_failure(port) [ 609.049505] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.049505] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.050225] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 609.050225] env[61991]: Removing descriptor: 18 [ 609.112676] env[61991]: DEBUG nova.scheduler.client.report [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.187833] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.213070] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.213959] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.213959] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.213959] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.213959] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.213959] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.214195] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.214283] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.214438] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.214589] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.214751] env[61991]: DEBUG nova.virt.hardware [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.215601] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12383060-e809-4401-8303-b31cd8a40c45 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.223278] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf758f4-9151-4ded-83fd-311564b6d44b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.236371] env[61991]: ERROR nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Traceback (most recent call last): [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] yield resources [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self.driver.spawn(context, instance, image_meta, [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] vm_ref = self.build_virtual_machine(instance, [ 609.236371] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] for vif in network_info: [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] return self._sync_wrapper(fn, *args, **kwargs) [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self.wait() [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self[:] = self._gt.wait() [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] return self._exit_event.wait() [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 609.236723] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] current.throw(*self._exc) [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] result = function(*args, **kwargs) [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] return func(*args, **kwargs) [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] raise e [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] nwinfo = self.network_api.allocate_for_instance( [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] created_port_ids = self._update_ports_for_instance( [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] with excutils.save_and_reraise_exception(): [ 609.237294] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self.force_reraise() [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] raise self.value [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] updated_port = self._update_port( [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] _ensure_no_port_binding_failure(port) [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] raise exception.PortBindingFailed(port_id=port['id']) [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 609.237634] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] [ 609.237634] env[61991]: INFO nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Terminating instance [ 609.239883] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.432737] env[61991]: DEBUG nova.network.neutron [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.493299] env[61991]: DEBUG nova.network.neutron [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.564703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.618416] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.619597] env[61991]: DEBUG nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.622719] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.057s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.995081] env[61991]: DEBUG oslo_concurrency.lockutils [req-082080b2-58ed-4091-a327-026283cdf997 req-c842a9d7-b7e6-4713-9a58-47cc85f1af61 service nova] Releasing lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.996801] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquired lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.996801] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.127452] env[61991]: DEBUG nova.compute.utils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.132363] env[61991]: DEBUG nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 610.485344] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba3b9ee-692c-4506-80ff-60fe68859021 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.493976] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1282049-309a-40e3-bbf7-a77764e48859 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.526928] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab606e0-2d88-4ce7-b654-7324954af1f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.534767] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c713314-1f48-4be7-a190-d695074c77bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.540079] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.549755] env[61991]: DEBUG nova.compute.provider_tree [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.633703] env[61991]: DEBUG nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.677668] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.936743] env[61991]: DEBUG nova.compute.manager [req-eb96a30f-3ef3-4c37-ad13-098d17449a6c req-9cb2a9b6-b7ff-42e1-ba3c-38a01f9bfd05 service nova] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Received event network-vif-deleted-702476ab-f514-40a7-ad38-44acef030059 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.052739] env[61991]: DEBUG nova.scheduler.client.report [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.179671] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Releasing lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.180287] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.180552] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.181257] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-086ba9e1-a462-481c-85cd-3d5a6022252d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.190651] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5962fd-b618-4c9c-a51d-aa05ab0d541b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.214477] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance acc99cda-1a6c-4c2b-a2e3-c92d646da723 could not be found. [ 611.214695] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.214880] env[61991]: INFO nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Took 0.03 seconds to destroy the instance on the hypervisor. [ 611.215129] env[61991]: DEBUG oslo.service.loopingcall [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.215331] env[61991]: DEBUG nova.compute.manager [-] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.215635] env[61991]: DEBUG nova.network.neutron [-] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.230338] env[61991]: DEBUG nova.network.neutron [-] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.557534] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.558814] env[61991]: ERROR nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Traceback (most recent call last): [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self.driver.spawn(context, instance, image_meta, [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] vm_ref = self.build_virtual_machine(instance, [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.558814] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] for vif in network_info: [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return self._sync_wrapper(fn, *args, **kwargs) [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self.wait() [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self[:] = self._gt.wait() [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return self._exit_event.wait() [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] result = hub.switch() [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.559220] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return self.greenlet.switch() [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] result = function(*args, **kwargs) [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] return func(*args, **kwargs) [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] raise e [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] nwinfo = self.network_api.allocate_for_instance( [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] created_port_ids = self._update_ports_for_instance( [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] with excutils.save_and_reraise_exception(): [ 611.559581] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] self.force_reraise() [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] raise self.value [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] updated_port = self._update_port( [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] _ensure_no_port_binding_failure(port) [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] raise exception.PortBindingFailed(port_id=port['id']) [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] nova.exception.PortBindingFailed: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. [ 611.559958] env[61991]: ERROR nova.compute.manager [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] [ 611.560288] env[61991]: DEBUG nova.compute.utils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.561025] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Build of instance 69db1562-7169-4ebc-899d-c54ee6e03384 was re-scheduled: Binding failed for port 2b3f9d6e-7ca9-4d27-b0e0-dcd657779c5a, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.561151] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.561363] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquiring lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.561594] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Acquired lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.561793] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.566023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.591s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.566023] env[61991]: INFO nova.compute.claims [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.645202] env[61991]: DEBUG nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.670734] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.671266] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.671266] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.671404] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.671528] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.671668] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.671865] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.675023] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.675023] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.675023] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.675023] env[61991]: DEBUG nova.virt.hardware [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.675023] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc588fb-148e-4faf-9a98-8c9810a97fa0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.681566] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9a3658-b5d7-49db-9096-83b2f28da5d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.694802] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 611.700131] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Creating folder: Project (3aa28f1a3b0e4842b8d22478fdd617ce). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 611.700392] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d35be38d-2fb4-4278-9143-4f8bb9a554b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.712270] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Created folder: Project (3aa28f1a3b0e4842b8d22478fdd617ce) in parent group-v246753. [ 611.712270] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Creating folder: Instances. Parent ref: group-v246761. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 611.712492] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67de2e30-869a-42b3-89b9-6a2e51d7dac8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.723967] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Created folder: Instances in parent group-v246761. [ 611.724233] env[61991]: DEBUG oslo.service.loopingcall [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.724414] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 611.724602] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13499bed-2639-4d36-84b0-1ed7fc58a6f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.736920] env[61991]: DEBUG nova.network.neutron [-] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.741767] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 611.741767] env[61991]: value = "task-1129280" [ 611.741767] env[61991]: _type = "Task" [ 611.741767] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.749526] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129280, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.080854] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.162935] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.239234] env[61991]: INFO nova.compute.manager [-] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Took 1.02 seconds to deallocate network for instance. [ 612.241643] env[61991]: DEBUG nova.compute.claims [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.242044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.251520] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129280, 'name': CreateVM_Task, 'duration_secs': 0.244863} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.252211] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 612.252840] env[61991]: DEBUG oslo_vmware.service [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27788cd9-c896-48ca-b4ec-0289aa632e8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.259228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.259402] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.259752] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.259991] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bc5c0e2-1051-4164-9680-c4573ad83376 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.265999] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 612.265999] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ef5c5a-5bef-230d-677c-08e1e8d0601a" [ 612.265999] env[61991]: _type = "Task" [ 612.265999] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.274339] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ef5c5a-5bef-230d-677c-08e1e8d0601a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.666559] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Releasing lock "refresh_cache-69db1562-7169-4ebc-899d-c54ee6e03384" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.667418] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.667739] env[61991]: DEBUG nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.667958] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.694989] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.778357] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.778573] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 612.778810] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.778953] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.779135] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 612.781493] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bd173d4-46dd-4ee4-98e8-f163edf74778 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.798977] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.799154] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 612.799923] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5644b3a6-d7fc-4c0a-a229-e38881643c53 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.809122] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f639a6b-0ab3-482b-ab93-0c69337a4d3f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.815954] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 612.815954] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e49ea6-173c-1bd5-15cc-679007049589" [ 612.815954] env[61991]: _type = "Task" [ 612.815954] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.823577] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e49ea6-173c-1bd5-15cc-679007049589, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.020851] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61b94a7-6d7f-452e-a2c5-8f9466d9b955 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.028931] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4cf7cc-5ffe-4686-a750-18dbed09203a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.060261] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0406ba93-d73f-47ee-86f8-9b6ef08b46b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.068304] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b6d75c-f517-419d-9fe4-b9242b68817c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.082340] env[61991]: DEBUG nova.compute.provider_tree [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.199020] env[61991]: DEBUG nova.network.neutron [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.325994] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Preparing fetch location {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 613.326317] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Creating directory with path [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.326513] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-898cad4d-308c-4763-ac51-6a9beba55643 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.354449] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Created directory with path [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.354648] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Fetch image to [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 613.354817] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Downloading image file data 254d700f-2f5a-49a3-8762-cec07162124a to [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk on the data store datastore1 {{(pid=61991) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 613.355642] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bd916d-b101-40a0-af95-55685416302f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.362694] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a2e26f-6587-4adf-be95-fa3b4c7f7bce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.371641] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f82d94-09b6-468a-85b1-45c5f96ff319 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.402319] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174613f0-166f-4e94-89dc-904ef988e634 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.408456] env[61991]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-cdf4fd37-5c3f-4e68-906c-c5e3be642cd0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.497920] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Downloading image file data 254d700f-2f5a-49a3-8762-cec07162124a to the data store datastore1 {{(pid=61991) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 613.558443] env[61991]: DEBUG oslo_vmware.rw_handles [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61991) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 613.628835] env[61991]: DEBUG nova.scheduler.client.report [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.704696] env[61991]: INFO nova.compute.manager [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] [instance: 69db1562-7169-4ebc-899d-c54ee6e03384] Took 1.04 seconds to deallocate network for instance. [ 614.134641] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.135214] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.138238] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.233s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.143609] env[61991]: INFO nova.compute.claims [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.220813] env[61991]: DEBUG oslo_vmware.rw_handles [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Completed reading data from the image iterator. {{(pid=61991) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 614.221082] env[61991]: DEBUG oslo_vmware.rw_handles [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 614.370599] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Downloaded image file data 254d700f-2f5a-49a3-8762-cec07162124a to vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk on the data store datastore1 {{(pid=61991) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 614.372784] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Caching image {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 614.373049] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Copying Virtual Disk [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk to [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 614.373351] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef01206e-d40d-4b0f-b2de-2907b6a8adf7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.384405] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 614.384405] env[61991]: value = "task-1129281" [ 614.384405] env[61991]: _type = "Task" [ 614.384405] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.392409] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129281, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.652296] env[61991]: DEBUG nova.compute.utils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.654115] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.654330] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 614.711554] env[61991]: DEBUG nova.policy [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b10ac4cad934e7ea793120929ad84ae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ddab761ff5f49aeb8a4b611cd9d7603', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.743535] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "30bd843a-9d86-4de1-b011-600e2a93bcb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.743789] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.746396] env[61991]: INFO nova.scheduler.client.report [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Deleted allocations for instance 69db1562-7169-4ebc-899d-c54ee6e03384 [ 614.895729] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129281, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.028104] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Successfully created port: 6ea9f583-188d-4df3-b9f6-c2a97b3208ac {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.161037] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.261022] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1d0558c8-7236-4ffe-bbd1-4c6e5da6c5fc tempest-AttachInterfacesUnderV243Test-1549108002 tempest-AttachInterfacesUnderV243Test-1549108002-project-member] Lock "69db1562-7169-4ebc-899d-c54ee6e03384" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.355s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.399590] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129281, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672286} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.399841] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Copied Virtual Disk [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk to [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 615.400010] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleting the datastore file [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a/tmp-sparse.vmdk {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 615.400256] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ee3d1e4-5487-4d4c-aee8-227eb508e34e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.409515] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 615.409515] env[61991]: value = "task-1129282" [ 615.409515] env[61991]: _type = "Task" [ 615.409515] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.420706] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.539515] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9eac33-80c0-4688-a8a9-db2d1f15a78d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.548195] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44664e96-a72e-4e92-9dbb-e87af4af2f64 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.582151] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0801f64e-33c1-48ec-b6cb-7f4cb88178be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.589830] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19992fe3-1ebf-468b-b906-061405bd5012 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.603654] env[61991]: DEBUG nova.compute.provider_tree [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.763849] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.919392] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024775} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.919678] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 615.919886] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Moving file from [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8/254d700f-2f5a-49a3-8762-cec07162124a to [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a. {{(pid=61991) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 615.920144] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-7e7e8d71-9773-43de-b302-c2eec352883a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.927583] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 615.927583] env[61991]: value = "task-1129283" [ 615.927583] env[61991]: _type = "Task" [ 615.927583] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.936164] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129283, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.967786] env[61991]: DEBUG nova.compute.manager [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Received event network-changed-6ea9f583-188d-4df3-b9f6-c2a97b3208ac {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.967991] env[61991]: DEBUG nova.compute.manager [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Refreshing instance network info cache due to event network-changed-6ea9f583-188d-4df3-b9f6-c2a97b3208ac. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 615.968215] env[61991]: DEBUG oslo_concurrency.lockutils [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] Acquiring lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.968357] env[61991]: DEBUG oslo_concurrency.lockutils [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] Acquired lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.968513] env[61991]: DEBUG nova.network.neutron [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Refreshing network info cache for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 616.110017] env[61991]: DEBUG nova.scheduler.client.report [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.123494] env[61991]: ERROR nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 616.123494] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.123494] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.123494] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.123494] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.123494] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.123494] env[61991]: ERROR nova.compute.manager raise self.value [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.123494] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 616.123494] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.123494] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 616.124268] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.124268] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 616.124268] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 616.124268] env[61991]: ERROR nova.compute.manager [ 616.124268] env[61991]: Traceback (most recent call last): [ 616.124268] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 616.124268] env[61991]: listener.cb(fileno) [ 616.124268] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.124268] env[61991]: result = function(*args, **kwargs) [ 616.124268] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.124268] env[61991]: return func(*args, **kwargs) [ 616.124268] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.124268] env[61991]: raise e [ 616.124268] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.124268] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 616.124268] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.124268] env[61991]: created_port_ids = self._update_ports_for_instance( [ 616.124268] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.124268] env[61991]: with excutils.save_and_reraise_exception(): [ 616.124268] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.124268] env[61991]: self.force_reraise() [ 616.124268] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.124268] env[61991]: raise self.value [ 616.124268] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.124268] env[61991]: updated_port = self._update_port( [ 616.124268] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.124268] env[61991]: _ensure_no_port_binding_failure(port) [ 616.124268] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.124268] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 616.126777] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 616.126777] env[61991]: Removing descriptor: 18 [ 616.173082] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.200620] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.200864] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.201035] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.201300] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.201448] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.201599] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.201802] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.201959] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.202138] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.202301] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.202470] env[61991]: DEBUG nova.virt.hardware [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.203349] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b09658-eb98-470c-9fe2-80f42ac027c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.211647] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5531fab6-843d-4396-878c-ade59b34d785 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.234925] env[61991]: ERROR nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Traceback (most recent call last): [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] yield resources [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self.driver.spawn(context, instance, image_meta, [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] vm_ref = self.build_virtual_machine(instance, [ 616.234925] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] for vif in network_info: [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] return self._sync_wrapper(fn, *args, **kwargs) [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self.wait() [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self[:] = self._gt.wait() [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] return self._exit_event.wait() [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 616.235448] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] current.throw(*self._exc) [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] result = function(*args, **kwargs) [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] return func(*args, **kwargs) [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] raise e [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] nwinfo = self.network_api.allocate_for_instance( [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] created_port_ids = self._update_ports_for_instance( [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] with excutils.save_and_reraise_exception(): [ 616.236052] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self.force_reraise() [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] raise self.value [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] updated_port = self._update_port( [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] _ensure_no_port_binding_failure(port) [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] raise exception.PortBindingFailed(port_id=port['id']) [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 616.236678] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] [ 616.236678] env[61991]: INFO nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Terminating instance [ 616.238188] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.297512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.440250] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129283, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023836} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.440250] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] File moved {{(pid=61991) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 616.440250] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Cleaning up location [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8 {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 616.440553] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleting the datastore file [datastore1] vmware_temp/7fcc2fe6-b4e8-403e-98c4-35580c868db8 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.441579] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39a734bd-30ba-4e40-8b5a-306ef69443e7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.447369] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 616.447369] env[61991]: value = "task-1129284" [ 616.447369] env[61991]: _type = "Task" [ 616.447369] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.456028] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129284, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.485803] env[61991]: DEBUG nova.network.neutron [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.582912] env[61991]: DEBUG nova.network.neutron [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.614995] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.615543] env[61991]: DEBUG nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.618280] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.703s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.619987] env[61991]: INFO nova.compute.claims [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.963939] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129284, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024032} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.963989] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 616.965470] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edf87c4f-37f6-4fa3-be7a-5e92ae313066 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.970695] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 616.970695] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286632f-c0f1-d6a2-31bf-01e98234efb2" [ 616.970695] env[61991]: _type = "Task" [ 616.970695] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.978806] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286632f-c0f1-d6a2-31bf-01e98234efb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.086225] env[61991]: DEBUG oslo_concurrency.lockutils [req-cd2560d9-4e06-4872-9df9-18eb1f812e61 req-6c53837e-a5c8-4249-8387-ead38b5579a8 service nova] Releasing lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.086655] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.086844] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 617.125017] env[61991]: DEBUG nova.compute.utils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.132432] env[61991]: DEBUG nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 617.481856] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286632f-c0f1-d6a2-31bf-01e98234efb2, 'name': SearchDatastore_Task, 'duration_secs': 0.010234} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.482141] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.482377] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 617.482613] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4bfdd119-4efa-44e1-8588-d56f12233880 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.489094] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 617.489094] env[61991]: value = "task-1129285" [ 617.489094] env[61991]: _type = "Task" [ 617.489094] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.496574] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129285, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.613332] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.633452] env[61991]: DEBUG nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.971651] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.997925] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129285, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.092425] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7265a7-bb67-46c0-9ed9-0748b12f204f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.100665] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cebf8e-b6b6-42aa-acda-287c7592341e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.133877] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee801dd3-463a-48c1-8586-aacd2a17fe21 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.137243] env[61991]: DEBUG nova.compute.manager [req-4ad60f06-18c0-45d6-b95c-75c1a32f1467 req-4d1c9015-275b-4369-abd8-167f093636f6 service nova] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Received event network-vif-deleted-6ea9f583-188d-4df3-b9f6-c2a97b3208ac {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.145999] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51eda214-6f6e-43d0-b93c-f2101b8aa73d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.159896] env[61991]: DEBUG nova.compute.provider_tree [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.474287] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Releasing lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.474723] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.474917] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.475237] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0e6de4c-f5a3-4a49-941b-2c6c27c14e27 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.484312] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399f53ba-42c3-4496-8627-6c67a6069d0d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.502320] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129285, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516563} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.502569] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 618.502776] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 618.503009] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92f33ff6-e358-4e0f-9b27-da835a1a1184 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.509140] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f9a4042-d9e4-4201-8386-18637c39795f could not be found. [ 618.509347] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.509525] env[61991]: INFO nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 618.509753] env[61991]: DEBUG oslo.service.loopingcall [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.510853] env[61991]: DEBUG nova.compute.manager [-] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.510957] env[61991]: DEBUG nova.network.neutron [-] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.512608] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 618.512608] env[61991]: value = "task-1129286" [ 618.512608] env[61991]: _type = "Task" [ 618.512608] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.523015] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.526302] env[61991]: DEBUG nova.network.neutron [-] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.651436] env[61991]: DEBUG nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.663026] env[61991]: DEBUG nova.scheduler.client.report [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.677415] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.677540] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.677701] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.677890] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.678046] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.678197] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.678399] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.678553] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.678710] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.678865] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.679041] env[61991]: DEBUG nova.virt.hardware [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.679854] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fd5eed-9b47-4cb6-bde7-e6c17be31a06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.688198] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9607504e-8bfb-4331-81e2-1ebbcd2b4117 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.701659] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 618.707125] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Creating folder: Project (a1d5c41302244406bf6963e7cb2d5e21). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 618.707421] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-124c5b50-96ed-4af2-b83d-6294623b1227 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.715902] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Created folder: Project (a1d5c41302244406bf6963e7cb2d5e21) in parent group-v246753. [ 618.716087] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Creating folder: Instances. Parent ref: group-v246764. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 618.716287] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6eeda320-3210-4548-9dc5-1873031a8edd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.724180] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Created folder: Instances in parent group-v246764. [ 618.724403] env[61991]: DEBUG oslo.service.loopingcall [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.724576] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 618.724753] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e934f64-7d6a-4dbe-9031-51cb04fc6787 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.740164] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 618.740164] env[61991]: value = "task-1129289" [ 618.740164] env[61991]: _type = "Task" [ 618.740164] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.747686] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129289, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.021879] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119852} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.026026] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 619.026026] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcacbb0b-21d1-4723-88e7-c02965a3c122 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.034605] env[61991]: DEBUG nova.network.neutron [-] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.043923] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 619.044980] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38270303-6fa2-446a-8035-bde80806e42c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.064761] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 619.064761] env[61991]: value = "task-1129290" [ 619.064761] env[61991]: _type = "Task" [ 619.064761] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.073391] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.168106] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.168534] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.172100] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.751s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.249872] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129289, 'name': CreateVM_Task, 'duration_secs': 0.501487} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.250067] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 619.250493] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.250649] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.250952] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 619.251999] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b73cc8ce-a477-4d7c-a06e-381cb5d36f64 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.256142] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 619.256142] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526a4ea6-d011-c378-2088-57971705e80a" [ 619.256142] env[61991]: _type = "Task" [ 619.256142] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.263715] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526a4ea6-d011-c378-2088-57971705e80a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.544957] env[61991]: INFO nova.compute.manager [-] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Took 1.03 seconds to deallocate network for instance. [ 619.547457] env[61991]: DEBUG nova.compute.claims [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.547639] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.576048] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129290, 'name': ReconfigVM_Task, 'duration_secs': 0.476262} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.576321] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Reconfigured VM instance instance-0000001a to attach disk [datastore1] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 619.576868] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-008ce495-9b0d-4fc8-bc6f-b94bdb00fa13 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.582883] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 619.582883] env[61991]: value = "task-1129291" [ 619.582883] env[61991]: _type = "Task" [ 619.582883] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.591476] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129291, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.679411] env[61991]: DEBUG nova.compute.utils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.684042] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.684042] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 619.770065] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526a4ea6-d011-c378-2088-57971705e80a, 'name': SearchDatastore_Task, 'duration_secs': 0.015891} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.771515] env[61991]: DEBUG nova.policy [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f680f8e6980641f6b01bb17379c2ac36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bc4cd1065084338924db8b72b397086', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.773066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.773299] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 619.773523] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.773661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.773839] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 619.776219] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77ebe625-c37a-43aa-8f69-7b65931cd90d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.784423] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 619.784597] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 619.785293] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-814d7374-18c0-430f-bbcb-0830b6f4a846 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.796949] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 619.796949] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c42643-5e7d-a00b-c367-96722357dae7" [ 619.796949] env[61991]: _type = "Task" [ 619.796949] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.806056] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c42643-5e7d-a00b-c367-96722357dae7, 'name': SearchDatastore_Task, 'duration_secs': 0.010338} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.806463] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b82d220-0b37-446c-a500-8731f82158b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.814605] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 619.814605] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bafa56-9fbd-624a-01b9-72e379f23d96" [ 619.814605] env[61991]: _type = "Task" [ 619.814605] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.822620] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bafa56-9fbd-624a-01b9-72e379f23d96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.097053] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129291, 'name': Rename_Task, 'duration_secs': 0.148847} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.097332] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 620.097565] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7a12c35-aae1-474d-87a2-107cc278c857 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.104282] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 620.104282] env[61991]: value = "task-1129292" [ 620.104282] env[61991]: _type = "Task" [ 620.104282] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.111791] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283b7cb3-3a0a-4bc7-a195-1ef232989c41 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.117564] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.121915] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6f2999-89a6-4623-a452-63c15e76431f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.156944] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027a95a6-0d47-4f07-9fcf-e576c742639e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.164389] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b46f7f0-ae80-41d3-9ad0-179fa0930994 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.177814] env[61991]: DEBUG nova.compute.provider_tree [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.185310] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.324742] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bafa56-9fbd-624a-01b9-72e379f23d96, 'name': SearchDatastore_Task, 'duration_secs': 0.009488} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.325017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.325276] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 620.325540] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c720f947-f8ec-40d9-9798-63296b151105 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.332887] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 620.332887] env[61991]: value = "task-1129293" [ 620.332887] env[61991]: _type = "Task" [ 620.332887] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.347421] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.374035] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Successfully created port: fb9df8fe-9922-4c59-a08c-693af5bba9ed {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.622747] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129292, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.681443] env[61991]: DEBUG nova.scheduler.client.report [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.691597] env[61991]: INFO nova.virt.block_device [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Booting with volume b991ee3d-75dd-4b90-90d5-4ecae72d7b45 at /dev/sda [ 620.763246] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e4c3dda-c4bf-43b4-890d-96275aaf2a24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.771374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43702a3f-df6d-4be2-b1b5-37a842f907bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.796429] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e614c5e-15af-49af-859f-170f2cd74137 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.806290] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4745746e-2cc2-4ef7-9873-5ed7fe35fb72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.831036] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a575d20-1727-4569-9395-6540fc409659 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.841913] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8ddaf6-b9b0-4da5-8acd-69ee766bc13e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.847634] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129293, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.861475] env[61991]: DEBUG nova.virt.block_device [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Updating existing volume attachment record: c03a33b9-6864-4fb0-976e-d6ffeb7cc68f {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 621.115975] env[61991]: DEBUG oslo_vmware.api [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129292, 'name': PowerOnVM_Task, 'duration_secs': 0.515532} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.116488] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 621.116686] env[61991]: INFO nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Took 9.47 seconds to spawn the instance on the hypervisor. [ 621.116864] env[61991]: DEBUG nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 621.117641] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c61e2d9-35d1-4714-b1c3-b8adc8712a07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.187710] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.188353] env[61991]: ERROR nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] Traceback (most recent call last): [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self.driver.spawn(context, instance, image_meta, [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] vm_ref = self.build_virtual_machine(instance, [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.188353] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] for vif in network_info: [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] return self._sync_wrapper(fn, *args, **kwargs) [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self.wait() [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self[:] = self._gt.wait() [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] return self._exit_event.wait() [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] current.throw(*self._exc) [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.188673] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] result = function(*args, **kwargs) [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] return func(*args, **kwargs) [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] raise e [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] nwinfo = self.network_api.allocate_for_instance( [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] created_port_ids = self._update_ports_for_instance( [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] with excutils.save_and_reraise_exception(): [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] self.force_reraise() [ 621.189048] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] raise self.value [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] updated_port = self._update_port( [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] _ensure_no_port_binding_failure(port) [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] raise exception.PortBindingFailed(port_id=port['id']) [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] nova.exception.PortBindingFailed: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. [ 621.189402] env[61991]: ERROR nova.compute.manager [instance: b7237462-348e-4e88-b415-a301499fab83] [ 621.189402] env[61991]: DEBUG nova.compute.utils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.193498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.382s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.193498] env[61991]: INFO nova.compute.claims [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.195843] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Build of instance b7237462-348e-4e88-b415-a301499fab83 was re-scheduled: Binding failed for port 0e78c05b-7e20-49ef-8118-cd1600aa4d23, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.196307] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.196533] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.196680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.196840] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.347488] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129293, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529199} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.347488] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.347488] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 621.347488] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2580f7d-0cbd-4f35-a82a-d76c06b1d96e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.353899] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 621.353899] env[61991]: value = "task-1129294" [ 621.353899] env[61991]: _type = "Task" [ 621.353899] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.362342] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129294, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.645126] env[61991]: INFO nova.compute.manager [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Took 30.66 seconds to build instance. [ 621.728751] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.863330] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129294, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.241187} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.863605] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.864437] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f44d7c-5ec4-435a-87e7-9615718aa3b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.885315] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.885567] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9e3e943-dd62-4ad1-b208-60fffc79731f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.908597] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 621.908597] env[61991]: value = "task-1129295" [ 621.908597] env[61991]: _type = "Task" [ 621.908597] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.917360] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.922316] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.063138] env[61991]: DEBUG nova.compute.manager [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] [instance: 32736198-a819-4afc-929b-a2c890263059] Received event network-changed-fb9df8fe-9922-4c59-a08c-693af5bba9ed {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.063498] env[61991]: DEBUG nova.compute.manager [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] [instance: 32736198-a819-4afc-929b-a2c890263059] Refreshing instance network info cache due to event network-changed-fb9df8fe-9922-4c59-a08c-693af5bba9ed. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.064282] env[61991]: DEBUG oslo_concurrency.lockutils [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] Acquiring lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.064646] env[61991]: DEBUG oslo_concurrency.lockutils [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] Acquired lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.064646] env[61991]: DEBUG nova.network.neutron [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] [instance: 32736198-a819-4afc-929b-a2c890263059] Refreshing network info cache for port fb9df8fe-9922-4c59-a08c-693af5bba9ed {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 622.147652] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0252652f-99cf-4aa4-aea1-09b4a74d3555 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "c2ef2717-4eac-428b-92a8-12a37d7c0719" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.495s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.187257] env[61991]: INFO nova.compute.manager [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Rebuilding instance [ 622.245913] env[61991]: DEBUG nova.compute.manager [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 622.246762] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e926268c-3eb4-455c-b1ca-fb8cc539c734 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.255354] env[61991]: ERROR nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 622.255354] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.255354] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.255354] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.255354] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.255354] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.255354] env[61991]: ERROR nova.compute.manager raise self.value [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.255354] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.255354] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.255354] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.255788] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.255788] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.255788] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 622.255788] env[61991]: ERROR nova.compute.manager [ 622.255788] env[61991]: Traceback (most recent call last): [ 622.255788] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.255788] env[61991]: listener.cb(fileno) [ 622.255788] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.255788] env[61991]: result = function(*args, **kwargs) [ 622.255788] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.255788] env[61991]: return func(*args, **kwargs) [ 622.255788] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.255788] env[61991]: raise e [ 622.255788] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.255788] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 622.255788] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.255788] env[61991]: created_port_ids = self._update_ports_for_instance( [ 622.255788] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.255788] env[61991]: with excutils.save_and_reraise_exception(): [ 622.255788] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.255788] env[61991]: self.force_reraise() [ 622.255788] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.255788] env[61991]: raise self.value [ 622.255788] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.255788] env[61991]: updated_port = self._update_port( [ 622.255788] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.255788] env[61991]: _ensure_no_port_binding_failure(port) [ 622.255788] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.255788] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.256517] env[61991]: nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 622.256517] env[61991]: Removing descriptor: 18 [ 622.418810] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129295, 'name': ReconfigVM_Task, 'duration_secs': 0.266945} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.421426] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.422209] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c83f206b-df78-4b68-9fea-9748a7983350 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.427986] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-b7237462-348e-4e88-b415-a301499fab83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.427986] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.427986] env[61991]: DEBUG nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.427986] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 622.431020] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 622.431020] env[61991]: value = "task-1129296" [ 622.431020] env[61991]: _type = "Task" [ 622.431020] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.441621] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129296, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.453673] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.590428] env[61991]: DEBUG nova.network.neutron [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.639296] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281f95db-7924-4a80-b5a3-7e91279cb75a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.646966] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd89ba26-d820-4699-92d1-3b2f89b3faa7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.651032] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.688939] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8253d67a-f2e1-4715-9b4b-76e2caed3031 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.698215] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64efca28-b13a-4c64-a46b-48514c38b563 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.701907] env[61991]: DEBUG nova.network.neutron [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] [instance: 32736198-a819-4afc-929b-a2c890263059] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.713026] env[61991]: DEBUG nova.compute.provider_tree [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.761478] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 622.761771] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d64de0fb-b81c-4e12-b6f6-db6fe52a1b1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.772602] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 622.772602] env[61991]: value = "task-1129297" [ 622.772602] env[61991]: _type = "Task" [ 622.772602] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.782493] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129297, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.943008] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129296, 'name': Rename_Task, 'duration_secs': 0.130314} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.943314] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 622.943570] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62d210de-acd5-4a5a-91f8-4b8acda3cc3e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.949797] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 622.949797] env[61991]: value = "task-1129298" [ 622.949797] env[61991]: _type = "Task" [ 622.949797] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.956601] env[61991]: DEBUG nova.network.neutron [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.957664] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129298, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.980787] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.984132] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.984132] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.984132] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.984132] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.984531] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.984963] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.985247] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.985603] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.985849] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.986171] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.986577] env[61991]: DEBUG nova.virt.hardware [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.987683] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c37719-6158-401b-a64b-8094e5bf54fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.998499] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cb379d-2153-408e-9f79-93c515f26ae4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.013070] env[61991]: ERROR nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] Traceback (most recent call last): [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] yield resources [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self.driver.spawn(context, instance, image_meta, [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] vm_ref = self.build_virtual_machine(instance, [ 623.013070] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] for vif in network_info: [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] return self._sync_wrapper(fn, *args, **kwargs) [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self.wait() [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self[:] = self._gt.wait() [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] return self._exit_event.wait() [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 623.013412] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] current.throw(*self._exc) [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] result = function(*args, **kwargs) [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] return func(*args, **kwargs) [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] raise e [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] nwinfo = self.network_api.allocate_for_instance( [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] created_port_ids = self._update_ports_for_instance( [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] with excutils.save_and_reraise_exception(): [ 623.013740] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self.force_reraise() [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] raise self.value [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] updated_port = self._update_port( [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] _ensure_no_port_binding_failure(port) [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] raise exception.PortBindingFailed(port_id=port['id']) [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 623.014088] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] [ 623.014088] env[61991]: INFO nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Terminating instance [ 623.016591] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquiring lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.174603] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.204412] env[61991]: DEBUG oslo_concurrency.lockutils [req-fb138688-b6a7-46e6-b3f4-e3f228fab99d req-078389e5-900f-4538-8532-6d05d9a8fde6 service nova] Releasing lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.205066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquired lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.205274] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.215833] env[61991]: DEBUG nova.scheduler.client.report [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.281363] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129297, 'name': PowerOffVM_Task, 'duration_secs': 0.130462} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.281604] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 623.281819] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 623.282557] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88efe45e-4922-4015-b768-5bd1b9b87ab8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.288856] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 623.289061] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-301a20ad-d962-4794-a99c-4df0aa253b96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.316712] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 623.317922] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 623.317922] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleting the datastore file [datastore1] c2ef2717-4eac-428b-92a8-12a37d7c0719 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 623.317922] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1da5cd36-4432-4b3e-a111-415d6da601e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.323546] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 623.323546] env[61991]: value = "task-1129300" [ 623.323546] env[61991]: _type = "Task" [ 623.323546] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.332018] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.458630] env[61991]: INFO nova.compute.manager [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: b7237462-348e-4e88-b415-a301499fab83] Took 1.03 seconds to deallocate network for instance. [ 623.461257] env[61991]: DEBUG oslo_vmware.api [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129298, 'name': PowerOnVM_Task, 'duration_secs': 0.468202} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.461687] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 623.461876] env[61991]: INFO nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Took 4.81 seconds to spawn the instance on the hypervisor. [ 623.462058] env[61991]: DEBUG nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 623.462823] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5db6d82-08d7-4b2e-a0ff-62ddfd90825a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.577564] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.577927] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.720483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.721057] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.723776] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.507s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.723950] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.724164] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 623.724471] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.840s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.727697] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b079389-79b3-45f5-80db-eb56525294fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.736189] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.739043] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aa06a9-c1aa-43e4-9d51-e30558d14cb2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.755944] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d424374-d20a-47f9-a33e-9669b0187896 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.763999] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cb8a08-0840-48a2-ae3b-f9fc01c342b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.796019] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181543MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 623.796198] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.834634] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213406} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.838024] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 623.838024] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 623.838024] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.872711] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.984939] env[61991]: INFO nova.compute.manager [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Took 26.10 seconds to build instance. [ 624.094688] env[61991]: DEBUG nova.compute.manager [req-ace29fc7-8500-481b-aba2-3324c6ac55f3 req-e4a61905-6010-4211-8471-589c35db120c service nova] [instance: 32736198-a819-4afc-929b-a2c890263059] Received event network-vif-deleted-fb9df8fe-9922-4c59-a08c-693af5bba9ed {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.228807] env[61991]: DEBUG nova.compute.utils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.230210] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.230370] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 624.274058] env[61991]: DEBUG nova.policy [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbb68bcfad394700b365bbd5bb60b098', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53dfdbbaa1094205b865d3fcf591136e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.301932] env[61991]: INFO nova.compute.manager [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Rebuilding instance [ 624.338428] env[61991]: DEBUG nova.compute.manager [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 624.339299] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a78e15-5386-46a2-acda-4cdaa972ad28 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.374764] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Releasing lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.375353] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.375964] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5c0d39d-0a21-4415-a696-2b1a427ac3f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.390587] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac463ff-cbe7-4db9-80de-48a10c6e31c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.416798] env[61991]: WARNING nova.virt.vmwareapi.driver [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 32736198-a819-4afc-929b-a2c890263059 could not be found. [ 624.417047] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.417337] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5492fc59-afc4-4677-ac8f-9b11790a9c4c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.428362] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f6175c-a43f-4216-aa22-a02dfd33d990 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.452307] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32736198-a819-4afc-929b-a2c890263059 could not be found. [ 624.452527] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.452701] env[61991]: INFO nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Took 0.08 seconds to destroy the instance on the hypervisor. [ 624.452940] env[61991]: DEBUG oslo.service.loopingcall [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.455355] env[61991]: DEBUG nova.compute.manager [-] [instance: 32736198-a819-4afc-929b-a2c890263059] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.455460] env[61991]: DEBUG nova.network.neutron [-] [instance: 32736198-a819-4afc-929b-a2c890263059] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.489328] env[61991]: DEBUG oslo_concurrency.lockutils [None req-60f28f37-d878-428e-aee3-fb7b304c50d7 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "432ab035-b5b0-4390-bef2-ec2691b4422b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.728s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.490711] env[61991]: INFO nova.scheduler.client.report [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted allocations for instance b7237462-348e-4e88-b415-a301499fab83 [ 624.496461] env[61991]: DEBUG nova.network.neutron [-] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.642110] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e84e028-a391-4a55-8c39-f7990c2a2367 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.649255] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42df595f-f9c7-476a-a09d-325aff6bc7e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.652806] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Successfully created port: 9d03dcc3-8f67-4c34-921f-d236df62870d {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.681279] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f970a0b-94b0-460c-b5ce-a1fd8caaee42 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.688517] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf467ad-3ba7-471a-a2e6-c25bfb8813da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.703164] env[61991]: DEBUG nova.compute.provider_tree [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.734044] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.854645] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 624.854931] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9cf508d-830e-446d-b4e8-f5e48244b5af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.864017] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 624.864017] env[61991]: value = "task-1129301" [ 624.864017] env[61991]: _type = "Task" [ 624.864017] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.872214] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.874827] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.875104] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.875157] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.875295] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.875438] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.875579] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.875777] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.875929] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.876102] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.876261] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.876429] env[61991]: DEBUG nova.virt.hardware [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.877214] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a0a095-0f41-4d8c-9ffb-f08c04b91dc4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.884914] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecec5cfb-313e-40e8-9d24-0facd138861f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.898533] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 624.904325] env[61991]: DEBUG oslo.service.loopingcall [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.905039] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 624.905408] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-289f606c-f6a1-45ad-b61b-6814ed143fdf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.923180] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 624.923180] env[61991]: value = "task-1129302" [ 624.923180] env[61991]: _type = "Task" [ 624.923180] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.930948] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129302, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.999660] env[61991]: DEBUG nova.network.neutron [-] [instance: 32736198-a819-4afc-929b-a2c890263059] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.001294] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.005458] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79051f3a-53d7-4158-b8af-84c69a6bc2be tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "b7237462-348e-4e88-b415-a301499fab83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.267s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.206422] env[61991]: DEBUG nova.scheduler.client.report [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.372410] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129301, 'name': PowerOffVM_Task, 'duration_secs': 0.188505} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.372670] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 625.372880] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.373668] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff47718b-e6ac-416a-ae16-2d6e854882ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.380762] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 625.380986] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2959d49-a52e-4089-858a-7a737a6f1319 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.410028] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 625.410201] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 625.410376] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Deleting the datastore file [datastore2] 432ab035-b5b0-4390-bef2-ec2691b4422b {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 625.411554] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1d5e57d-c60b-4e93-be4b-991afa6581d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.418184] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 625.418184] env[61991]: value = "task-1129304" [ 625.418184] env[61991]: _type = "Task" [ 625.418184] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.426171] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.433285] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129302, 'name': CreateVM_Task, 'duration_secs': 0.246007} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.433453] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 625.433848] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.434008] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.434350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 625.434586] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98943fe6-1cd4-4bc4-ba26-75b07595bde4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.438539] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 625.438539] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52109715-4140-b806-9f13-be99589edef0" [ 625.438539] env[61991]: _type = "Task" [ 625.438539] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.449247] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52109715-4140-b806-9f13-be99589edef0, 'name': SearchDatastore_Task, 'duration_secs': 0.007821} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.449524] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.449746] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.449964] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.450117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.450283] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.450519] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-922f049d-3b88-41e7-9260-052a4657d81e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.457036] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.457260] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.457881] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6cafba2-8967-42a4-b283-1a7205afccfe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.462606] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 625.462606] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52881299-b2aa-5dbb-60ff-03bb27d4b926" [ 625.462606] env[61991]: _type = "Task" [ 625.462606] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.469856] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52881299-b2aa-5dbb-60ff-03bb27d4b926, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.508951] env[61991]: INFO nova.compute.manager [-] [instance: 32736198-a819-4afc-929b-a2c890263059] Took 1.05 seconds to deallocate network for instance. [ 625.511510] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.525795] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.711736] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.712386] env[61991]: ERROR nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Traceback (most recent call last): [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self.driver.spawn(context, instance, image_meta, [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] vm_ref = self.build_virtual_machine(instance, [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.712386] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] for vif in network_info: [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] return self._sync_wrapper(fn, *args, **kwargs) [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self.wait() [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self[:] = self._gt.wait() [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] return self._exit_event.wait() [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] current.throw(*self._exc) [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.712675] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] result = function(*args, **kwargs) [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] return func(*args, **kwargs) [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] raise e [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] nwinfo = self.network_api.allocate_for_instance( [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] created_port_ids = self._update_ports_for_instance( [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] with excutils.save_and_reraise_exception(): [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] self.force_reraise() [ 625.712981] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] raise self.value [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] updated_port = self._update_port( [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] _ensure_no_port_binding_failure(port) [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] raise exception.PortBindingFailed(port_id=port['id']) [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] nova.exception.PortBindingFailed: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. [ 625.713339] env[61991]: ERROR nova.compute.manager [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] [ 625.713339] env[61991]: DEBUG nova.compute.utils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.714305] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.150s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.715824] env[61991]: INFO nova.compute.claims [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.718395] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Build of instance 22be69a5-6405-4e8c-8cdd-0843847f9228 was re-scheduled: Binding failed for port f1e8449e-af0e-44f4-a850-8f4efe4b1e40, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.719174] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.719174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.719174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.719319] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.724460] env[61991]: ERROR nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 625.724460] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.724460] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.724460] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.724460] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.724460] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.724460] env[61991]: ERROR nova.compute.manager raise self.value [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.724460] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.724460] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.724460] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.724848] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.724848] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.724848] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 625.724848] env[61991]: ERROR nova.compute.manager [ 625.724848] env[61991]: Traceback (most recent call last): [ 625.724848] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.724848] env[61991]: listener.cb(fileno) [ 625.724848] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.724848] env[61991]: result = function(*args, **kwargs) [ 625.724848] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.724848] env[61991]: return func(*args, **kwargs) [ 625.724848] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.724848] env[61991]: raise e [ 625.724848] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.724848] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 625.724848] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.724848] env[61991]: created_port_ids = self._update_ports_for_instance( [ 625.724848] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.724848] env[61991]: with excutils.save_and_reraise_exception(): [ 625.724848] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.724848] env[61991]: self.force_reraise() [ 625.724848] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.724848] env[61991]: raise self.value [ 625.724848] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.724848] env[61991]: updated_port = self._update_port( [ 625.724848] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.724848] env[61991]: _ensure_no_port_binding_failure(port) [ 625.724848] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.724848] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.725533] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 625.725533] env[61991]: Removing descriptor: 18 [ 625.743863] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 625.767888] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.768142] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.768298] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.768479] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.768622] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.768823] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.768961] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.769128] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.769291] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.769448] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.769614] env[61991]: DEBUG nova.virt.hardware [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.770474] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8a98df-d1ed-4cb3-bc94-6402e1bbd1f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.778339] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f8d880-86c8-4037-8631-0ddd79544d61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.791854] env[61991]: ERROR nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Traceback (most recent call last): [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] yield resources [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self.driver.spawn(context, instance, image_meta, [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] vm_ref = self.build_virtual_machine(instance, [ 625.791854] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] for vif in network_info: [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] return self._sync_wrapper(fn, *args, **kwargs) [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self.wait() [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self[:] = self._gt.wait() [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] return self._exit_event.wait() [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 625.792415] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] current.throw(*self._exc) [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] result = function(*args, **kwargs) [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] return func(*args, **kwargs) [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] raise e [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] nwinfo = self.network_api.allocate_for_instance( [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] created_port_ids = self._update_ports_for_instance( [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] with excutils.save_and_reraise_exception(): [ 625.792785] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self.force_reraise() [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] raise self.value [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] updated_port = self._update_port( [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] _ensure_no_port_binding_failure(port) [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] raise exception.PortBindingFailed(port_id=port['id']) [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 625.793306] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] [ 625.793306] env[61991]: INFO nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Terminating instance [ 625.794242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.794416] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquired lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.794578] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.929999] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091402} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.930314] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.930500] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 625.930672] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.973797] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52881299-b2aa-5dbb-60ff-03bb27d4b926, 'name': SearchDatastore_Task, 'duration_secs': 0.008292} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.974629] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b4ca880-8fa9-4aef-9617-fb446e8b98c9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.979593] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 625.979593] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f08216-1e6b-a890-1757-fc4a08be1b34" [ 625.979593] env[61991]: _type = "Task" [ 625.979593] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.987554] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f08216-1e6b-a890-1757-fc4a08be1b34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.035022] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.078536] env[61991]: INFO nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Took 0.56 seconds to detach 1 volumes for instance. [ 626.080948] env[61991]: DEBUG nova.compute.claims [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.081174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.117425] env[61991]: DEBUG nova.compute.manager [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Received event network-changed-9d03dcc3-8f67-4c34-921f-d236df62870d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.117699] env[61991]: DEBUG nova.compute.manager [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Refreshing instance network info cache due to event network-changed-9d03dcc3-8f67-4c34-921f-d236df62870d. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 626.117808] env[61991]: DEBUG oslo_concurrency.lockutils [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] Acquiring lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.245792] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.321636] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.360336] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.424251] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.492568] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f08216-1e6b-a890-1757-fc4a08be1b34, 'name': SearchDatastore_Task, 'duration_secs': 0.008436} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.492828] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.493100] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 626.493377] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32a96d79-a23f-493e-b25e-829363a9eb75 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.501238] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 626.501238] env[61991]: value = "task-1129305" [ 626.501238] env[61991]: _type = "Task" [ 626.501238] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.508635] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.863785] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-22be69a5-6405-4e8c-8cdd-0843847f9228" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.864766] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.864766] env[61991]: DEBUG nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.864766] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.882367] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.927628] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Releasing lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.927895] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.928137] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.931154] env[61991]: DEBUG oslo_concurrency.lockutils [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] Acquired lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.931372] env[61991]: DEBUG nova.network.neutron [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Refreshing network info cache for port 9d03dcc3-8f67-4c34-921f-d236df62870d {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 626.933047] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-054c27ae-ae9c-4db7-aa64-4f5e71930479 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.942290] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093813a8-6c42-4ee1-aa48-b5a3f7542080 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.972385] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f720a778-5f29-4f27-8b84-7bf377f27dc9 could not be found. [ 626.972664] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.972845] env[61991]: INFO nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.973175] env[61991]: DEBUG oslo.service.loopingcall [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.975816] env[61991]: DEBUG nova.compute.manager [-] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.975919] env[61991]: DEBUG nova.network.neutron [-] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.989252] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.989252] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.989252] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.989252] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.989442] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.989442] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.989442] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.989717] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.989904] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.990120] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.990307] env[61991]: DEBUG nova.virt.hardware [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.991466] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57092848-c7da-4bf7-be40-295c7c1159dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.994693] env[61991]: DEBUG nova.network.neutron [-] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.008247] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad01d3f5-4ee4-43ee-b842-d278774b9461 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.020417] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129305, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496317} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.029039] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 627.029282] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.029860] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 627.039017] env[61991]: DEBUG oslo.service.loopingcall [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.039017] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93a165a6-4df2-482f-960b-ade14ccfe062 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.039956] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 627.041190] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7af51e3-c388-4085-b9fb-06c994c84fc1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.060172] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 627.060172] env[61991]: value = "task-1129306" [ 627.060172] env[61991]: _type = "Task" [ 627.060172] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.060172] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 627.060172] env[61991]: value = "task-1129307" [ 627.060172] env[61991]: _type = "Task" [ 627.060172] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.073846] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129306, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.076899] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129307, 'name': CreateVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.244499] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c95f127-15e5-44aa-8f3b-db8740a5ec04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.254279] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8173e99c-6384-4c82-bd6b-5659bc20b881 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.287570] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293041a0-a193-4616-b43f-f7c3ff878b27 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.296109] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c3a78b-375c-4582-8ba2-3ee0bf983282 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.309217] env[61991]: DEBUG nova.compute.provider_tree [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.385968] env[61991]: DEBUG nova.network.neutron [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.452950] env[61991]: DEBUG nova.network.neutron [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.499960] env[61991]: DEBUG nova.network.neutron [-] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.549318] env[61991]: DEBUG nova.network.neutron [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.574731] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129306, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068977} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.579157] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 627.579745] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129307, 'name': CreateVM_Task, 'duration_secs': 0.307447} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.580473] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c97736-f175-425a-9e43-bf448a5b869f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.582825] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.583268] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.583476] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.583739] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 627.584359] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96163fdf-1927-4dee-8f1e-5ec9d73e1a69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.602537] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 627.603672] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb811c64-39ea-4243-9a67-477914cf9207 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.619835] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 627.619835] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ce0657-1a37-7162-9101-8a60dd0c4b63" [ 627.619835] env[61991]: _type = "Task" [ 627.619835] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.626444] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 627.626444] env[61991]: value = "task-1129308" [ 627.626444] env[61991]: _type = "Task" [ 627.626444] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.629911] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ce0657-1a37-7162-9101-8a60dd0c4b63, 'name': SearchDatastore_Task, 'duration_secs': 0.009787} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.634131] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.634131] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.634131] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.634131] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.634302] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.634302] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97a66f61-9cb0-4936-856a-2e78c55b4e5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.645468] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129308, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.654987] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.655207] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 627.655979] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8898135c-70d3-48dc-9bf3-1397ada859ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.660943] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 627.660943] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c2af0b-e594-b96c-1a91-4cfc3b51979e" [ 627.660943] env[61991]: _type = "Task" [ 627.660943] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.668217] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c2af0b-e594-b96c-1a91-4cfc3b51979e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.812048] env[61991]: DEBUG nova.scheduler.client.report [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.892496] env[61991]: INFO nova.compute.manager [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 22be69a5-6405-4e8c-8cdd-0843847f9228] Took 1.03 seconds to deallocate network for instance. [ 628.004145] env[61991]: INFO nova.compute.manager [-] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Took 1.03 seconds to deallocate network for instance. [ 628.007911] env[61991]: DEBUG nova.compute.claims [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.007981] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.051883] env[61991]: DEBUG oslo_concurrency.lockutils [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] Releasing lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.052143] env[61991]: DEBUG nova.compute.manager [req-c60b5366-a7c6-473a-b0d5-6a8c7995c3a6 req-32f469e0-4a00-47ff-bc91-c693df16249a service nova] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Received event network-vif-deleted-9d03dcc3-8f67-4c34-921f-d236df62870d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.055008] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.055227] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "80d05278-b516-4408-94b0-11bc93500b5c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.139674] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129308, 'name': ReconfigVM_Task, 'duration_secs': 0.270699} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.139944] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Reconfigured VM instance instance-0000001a to attach disk [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 628.140553] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3be4da19-42a7-468d-84a9-a25fbcf621ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.146062] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 628.146062] env[61991]: value = "task-1129309" [ 628.146062] env[61991]: _type = "Task" [ 628.146062] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.153353] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129309, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.169089] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c2af0b-e594-b96c-1a91-4cfc3b51979e, 'name': SearchDatastore_Task, 'duration_secs': 0.009244} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.169830] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6870c27-fbf2-46d1-a8b4-d634ae21af32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.174964] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 628.174964] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bec0fb-f87e-4153-5af3-3a4d4adb7a08" [ 628.174964] env[61991]: _type = "Task" [ 628.174964] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.181523] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bec0fb-f87e-4153-5af3-3a4d4adb7a08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.316866] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.317437] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.320125] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.078s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.657629] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129309, 'name': Rename_Task, 'duration_secs': 0.129277} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.657910] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 628.658295] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4e48e80-1590-4096-bdb8-05771a2e7113 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.664663] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 628.664663] env[61991]: value = "task-1129310" [ 628.664663] env[61991]: _type = "Task" [ 628.664663] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.672324] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.683184] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bec0fb-f87e-4153-5af3-3a4d4adb7a08, 'name': SearchDatastore_Task, 'duration_secs': 0.00906} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.683433] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.683684] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 628.683916] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aabb4301-2e8b-4ba7-b1d0-035349f3e99e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.690479] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 628.690479] env[61991]: value = "task-1129311" [ 628.690479] env[61991]: _type = "Task" [ 628.690479] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.698464] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129311, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.824726] env[61991]: DEBUG nova.compute.utils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.831048] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.831278] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.896494] env[61991]: DEBUG nova.policy [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62bac57001954cae953f0f369ec98aa9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '61e5b67397f941a6b712e4b2e0c7a217', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.921791] env[61991]: INFO nova.scheduler.client.report [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocations for instance 22be69a5-6405-4e8c-8cdd-0843847f9228 [ 629.178920] env[61991]: DEBUG oslo_vmware.api [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129310, 'name': PowerOnVM_Task, 'duration_secs': 0.430443} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.179234] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 629.179505] env[61991]: DEBUG nova.compute.manager [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 629.180614] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05d4586-b370-4b2d-b479-23e0b73833a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.202982] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129311, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.207624] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Successfully created port: 409facee-38ec-442f-9b89-3afc939a8e18 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.253556] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13dbb18e-f9f6-42d1-8552-f074def4c4ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.262816] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150a1ba8-5eee-4ff2-be6f-96488069baeb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.296300] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd296373-5cd1-4292-a055-8e5d09116bce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.306356] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780b0d86-9ac0-4f1f-8e9e-2a1aa9d8b2e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.320708] env[61991]: DEBUG nova.compute.provider_tree [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.331489] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.431878] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c2346f48-6566-48c5-821d-ad68a2540517 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "22be69a5-6405-4e8c-8cdd-0843847f9228" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.070s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.703915] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.707622] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129311, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637059} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.707860] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 629.708079] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.708316] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b9937f5-7984-4a01-82ed-bc5217f52b10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.715675] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 629.715675] env[61991]: value = "task-1129312" [ 629.715675] env[61991]: _type = "Task" [ 629.715675] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.723798] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129312, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.824178] env[61991]: DEBUG nova.scheduler.client.report [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.936535] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.225732] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129312, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107009} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.226033] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 630.226826] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3b885c-b7fe-4ed3-92e7-f10c86cc86f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.248631] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.248915] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68be922e-573b-4509-b237-d6d9b11c1b6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.270278] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 630.270278] env[61991]: value = "task-1129313" [ 630.270278] env[61991]: _type = "Task" [ 630.270278] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.279154] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129313, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.329853] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.330534] env[61991]: ERROR nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Traceback (most recent call last): [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self.driver.spawn(context, instance, image_meta, [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] vm_ref = self.build_virtual_machine(instance, [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.330534] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] for vif in network_info: [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] return self._sync_wrapper(fn, *args, **kwargs) [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self.wait() [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self[:] = self._gt.wait() [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] return self._exit_event.wait() [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] current.throw(*self._exc) [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.330862] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] result = function(*args, **kwargs) [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] return func(*args, **kwargs) [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] raise e [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] nwinfo = self.network_api.allocate_for_instance( [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] created_port_ids = self._update_ports_for_instance( [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] with excutils.save_and_reraise_exception(): [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] self.force_reraise() [ 630.331250] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] raise self.value [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] updated_port = self._update_port( [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] _ensure_no_port_binding_failure(port) [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] raise exception.PortBindingFailed(port_id=port['id']) [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] nova.exception.PortBindingFailed: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. [ 630.331553] env[61991]: ERROR nova.compute.manager [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] [ 630.331553] env[61991]: DEBUG nova.compute.utils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 630.336903] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.035s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.336903] env[61991]: INFO nova.compute.claims [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.337105] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Build of instance acc99cda-1a6c-4c2b-a2e3-c92d646da723 was re-scheduled: Binding failed for port 702476ab-f514-40a7-ad38-44acef030059, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 630.341023] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 630.341023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquiring lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.341023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Acquired lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.341023] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.343055] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.361971] env[61991]: DEBUG nova.compute.manager [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Received event network-changed-409facee-38ec-442f-9b89-3afc939a8e18 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.362197] env[61991]: DEBUG nova.compute.manager [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Refreshing instance network info cache due to event network-changed-409facee-38ec-442f-9b89-3afc939a8e18. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 630.362438] env[61991]: DEBUG oslo_concurrency.lockutils [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] Acquiring lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.362585] env[61991]: DEBUG oslo_concurrency.lockutils [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] Acquired lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.362739] env[61991]: DEBUG nova.network.neutron [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Refreshing network info cache for port 409facee-38ec-442f-9b89-3afc939a8e18 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 630.372947] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.373198] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.373367] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.373549] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.373691] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.373836] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.374055] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.374244] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.374425] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.374590] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.374751] env[61991]: DEBUG nova.virt.hardware [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.375823] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0090f06f-3876-4bb7-bdf0-3d9b246e3f86 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.385666] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c50dab1-3573-4236-b18d-7a0b8b6cf84c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.458701] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.556260] env[61991]: INFO nova.compute.manager [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Rebuilding instance [ 630.597530] env[61991]: ERROR nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 630.597530] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.597530] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.597530] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.597530] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.597530] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.597530] env[61991]: ERROR nova.compute.manager raise self.value [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.597530] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.597530] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.597530] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.598372] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.598372] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.598372] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 630.598372] env[61991]: ERROR nova.compute.manager [ 630.598372] env[61991]: Traceback (most recent call last): [ 630.598372] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.598372] env[61991]: listener.cb(fileno) [ 630.598372] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.598372] env[61991]: result = function(*args, **kwargs) [ 630.598372] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.598372] env[61991]: return func(*args, **kwargs) [ 630.598372] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.598372] env[61991]: raise e [ 630.598372] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.598372] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 630.598372] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.598372] env[61991]: created_port_ids = self._update_ports_for_instance( [ 630.598372] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.598372] env[61991]: with excutils.save_and_reraise_exception(): [ 630.598372] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.598372] env[61991]: self.force_reraise() [ 630.598372] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.598372] env[61991]: raise self.value [ 630.598372] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.598372] env[61991]: updated_port = self._update_port( [ 630.598372] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.598372] env[61991]: _ensure_no_port_binding_failure(port) [ 630.598372] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.598372] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.599205] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 630.599205] env[61991]: Removing descriptor: 17 [ 630.599205] env[61991]: ERROR nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Traceback (most recent call last): [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] yield resources [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self.driver.spawn(context, instance, image_meta, [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.599205] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] vm_ref = self.build_virtual_machine(instance, [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] for vif in network_info: [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return self._sync_wrapper(fn, *args, **kwargs) [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self.wait() [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self[:] = self._gt.wait() [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return self._exit_event.wait() [ 630.599581] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] result = hub.switch() [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return self.greenlet.switch() [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] result = function(*args, **kwargs) [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return func(*args, **kwargs) [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] raise e [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] nwinfo = self.network_api.allocate_for_instance( [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.600379] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] created_port_ids = self._update_ports_for_instance( [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] with excutils.save_and_reraise_exception(): [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self.force_reraise() [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] raise self.value [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] updated_port = self._update_port( [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] _ensure_no_port_binding_failure(port) [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.600964] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] raise exception.PortBindingFailed(port_id=port['id']) [ 630.601554] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 630.601554] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] [ 630.601554] env[61991]: INFO nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Terminating instance [ 630.605853] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.615962] env[61991]: DEBUG nova.compute.manager [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 630.616799] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5608c9-a73f-4051-8be3-5ae176661087 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.779365] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129313, 'name': ReconfigVM_Task, 'duration_secs': 0.275992} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.779666] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 432ab035-b5b0-4390-bef2-ec2691b4422b/432ab035-b5b0-4390-bef2-ec2691b4422b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 630.780351] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-895a36ad-dd76-4d83-8990-00f3d4e2eebe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.786831] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 630.786831] env[61991]: value = "task-1129314" [ 630.786831] env[61991]: _type = "Task" [ 630.786831] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.795372] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129314, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.859808] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.886790] env[61991]: DEBUG nova.network.neutron [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.966495] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.005195] env[61991]: DEBUG nova.network.neutron [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.128102] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 631.128313] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9912f604-3d58-4a7d-9968-3e7b8b486876 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.136160] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 631.136160] env[61991]: value = "task-1129315" [ 631.136160] env[61991]: _type = "Task" [ 631.136160] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.146743] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.296445] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129314, 'name': Rename_Task, 'duration_secs': 0.150871} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.296732] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 631.297097] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce734e10-8afb-4454-8882-fd8733b835e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.302863] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 631.302863] env[61991]: value = "task-1129316" [ 631.302863] env[61991]: _type = "Task" [ 631.302863] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.309999] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129316, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.473025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Releasing lock "refresh_cache-acc99cda-1a6c-4c2b-a2e3-c92d646da723" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.473025] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.473025] env[61991]: DEBUG nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.473025] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.482565] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "78631e83-e574-4a90-b062-176b2d12d681" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.482734] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.501947] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.506916] env[61991]: DEBUG oslo_concurrency.lockutils [req-93632ecf-cb38-4151-a41a-2fdf849a0caf req-9662a692-32cc-441d-b2dd-ae5c14d6da50 service nova] Releasing lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.507523] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquired lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.507523] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.648647] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129315, 'name': PowerOffVM_Task, 'duration_secs': 0.132563} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.651119] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 631.651363] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.652270] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9848de3-4f76-483d-8364-e7d67dfe91af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.660972] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 631.663951] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f662237a-2a44-4e10-93a5-73d2240f3171 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.692264] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 631.692492] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 631.692669] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Deleting the datastore file [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 631.692926] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18f8c963-855f-4e00-bd45-ac67db471ba8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.698881] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 631.698881] env[61991]: value = "task-1129318" [ 631.698881] env[61991]: _type = "Task" [ 631.698881] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.709430] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.781702] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0287cb5e-1e50-4499-8e92-41decb6e77ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.789110] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7796e1f-d323-4663-872b-a2ae4d6305ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.822400] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629c777e-aae3-464c-8d85-661d5eee2944 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.829950] env[61991]: DEBUG oslo_vmware.api [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129316, 'name': PowerOnVM_Task, 'duration_secs': 0.416671} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.832023] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 631.832287] env[61991]: DEBUG nova.compute.manager [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 631.833064] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60ce19d-5169-4f52-a143-c12c535ffdc9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.836426] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3901e94-3179-4a1a-a573-a04f386a32dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.854574] env[61991]: DEBUG nova.compute.provider_tree [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.005026] env[61991]: DEBUG nova.network.neutron [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.027427] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.116544] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.208846] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280271} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.209132] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 632.209317] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 632.209493] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.361078] env[61991]: DEBUG nova.scheduler.client.report [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.365719] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.507598] env[61991]: INFO nova.compute.manager [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] [instance: acc99cda-1a6c-4c2b-a2e3-c92d646da723] Took 1.03 seconds to deallocate network for instance. [ 632.537466] env[61991]: DEBUG nova.compute.manager [req-61987a03-1b81-4a12-9705-7a1ab2ce4336 req-77529b23-f6ab-47e3-8de2-cc3700909dfd service nova] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Received event network-vif-deleted-409facee-38ec-442f-9b89-3afc939a8e18 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.618984] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Releasing lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.619474] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.619672] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.619960] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ffc1e50-0469-4c66-8a88-8db9ab8e6140 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.629814] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7752aac3-1baf-40ca-800b-253e469b704d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.655947] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 931d58eb-3bbc-4d81-b737-95830b342027 could not be found. [ 632.656192] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.656369] env[61991]: INFO nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.656604] env[61991]: DEBUG oslo.service.loopingcall [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.657376] env[61991]: DEBUG nova.compute.manager [-] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.657478] env[61991]: DEBUG nova.network.neutron [-] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 632.677805] env[61991]: DEBUG nova.network.neutron [-] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.866744] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.867325] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 632.870593] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.323s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.143111] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "432ab035-b5b0-4390-bef2-ec2691b4422b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.143475] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "432ab035-b5b0-4390-bef2-ec2691b4422b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.143700] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "432ab035-b5b0-4390-bef2-ec2691b4422b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.143882] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "432ab035-b5b0-4390-bef2-ec2691b4422b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.144061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "432ab035-b5b0-4390-bef2-ec2691b4422b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.146302] env[61991]: INFO nova.compute.manager [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Terminating instance [ 633.147877] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "refresh_cache-432ab035-b5b0-4390-bef2-ec2691b4422b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.148044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquired lock "refresh_cache-432ab035-b5b0-4390-bef2-ec2691b4422b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.148211] env[61991]: DEBUG nova.network.neutron [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.181788] env[61991]: DEBUG nova.network.neutron [-] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.240168] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.240422] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.240581] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.240849] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.240921] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.241071] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.241313] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.241492] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.241662] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.241820] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.241991] env[61991]: DEBUG nova.virt.hardware [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.242888] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ec0dfb-cae1-4dbb-8816-395d47a50224 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.251588] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e092beeb-8725-41c2-b16b-21ccb7cee030 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.265093] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 633.270510] env[61991]: DEBUG oslo.service.loopingcall [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.270757] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 633.270963] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62f82fd1-4b9f-4df0-b834-060b336c1daa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.287020] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 633.287020] env[61991]: value = "task-1129319" [ 633.287020] env[61991]: _type = "Task" [ 633.287020] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.295782] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129319, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.372119] env[61991]: DEBUG nova.compute.utils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.373630] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.373797] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.442816] env[61991]: DEBUG nova.policy [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62bac57001954cae953f0f369ec98aa9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '61e5b67397f941a6b712e4b2e0c7a217', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.534058] env[61991]: INFO nova.scheduler.client.report [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Deleted allocations for instance acc99cda-1a6c-4c2b-a2e3-c92d646da723 [ 633.675121] env[61991]: DEBUG nova.network.neutron [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.684308] env[61991]: INFO nova.compute.manager [-] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Took 1.03 seconds to deallocate network for instance. [ 633.685829] env[61991]: DEBUG nova.compute.claims [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.685997] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.749052] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e87e3c-36e2-4d3c-92ab-bbc27f1dfc75 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.758214] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40333720-1d22-4b6e-adca-77ac945e903c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.766310] env[61991]: DEBUG nova.network.neutron [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.792382] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Successfully created port: b0a036d2-bd1d-4daa-9b89-cd85e219c8b5 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.799066] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b5a901-3ab8-4104-8f64-f6438af77339 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.810393] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13919305-e71f-44b1-af9f-a4ca61269168 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.814159] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129319, 'name': CreateVM_Task, 'duration_secs': 0.245558} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.814618] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 633.815360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.815566] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.815946] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 633.816222] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5164e05b-1888-437d-936b-0e363fdec785 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.825470] env[61991]: DEBUG nova.compute.provider_tree [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.830216] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 633.830216] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e723c6-3366-5d64-5a5e-ae8dc3c27f10" [ 633.830216] env[61991]: _type = "Task" [ 633.830216] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.839220] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e723c6-3366-5d64-5a5e-ae8dc3c27f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.877460] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.043560] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f48039c1-73b1-4ab2-93d7-069d90bb158b tempest-DeleteServersAdminTestJSON-1032479071 tempest-DeleteServersAdminTestJSON-1032479071-project-member] Lock "acc99cda-1a6c-4c2b-a2e3-c92d646da723" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.174s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.269030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Releasing lock "refresh_cache-432ab035-b5b0-4390-bef2-ec2691b4422b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.269030] env[61991]: DEBUG nova.compute.manager [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.269030] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.269030] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63840325-2f1f-48d9-80e2-bcb6811bdf87 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.276780] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 634.276966] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c2b9a33f-ff5f-4b26-aecb-720f8047a1e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.283446] env[61991]: DEBUG oslo_vmware.api [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 634.283446] env[61991]: value = "task-1129320" [ 634.283446] env[61991]: _type = "Task" [ 634.283446] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.294018] env[61991]: DEBUG oslo_vmware.api [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129320, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.328744] env[61991]: DEBUG nova.scheduler.client.report [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.341609] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e723c6-3366-5d64-5a5e-ae8dc3c27f10, 'name': SearchDatastore_Task, 'duration_secs': 0.008808} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.343942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.343942] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 634.343942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.343942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.344191] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 634.344191] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5832de92-6c69-40d1-a4c6-b770e5fbf741 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.350816] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 634.351014] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 634.351738] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-313c75a8-0720-4d06-a970-97b8d74fcd40 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.358208] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 634.358208] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c92c93-9652-dadd-23bb-6a60fa2bdc2b" [ 634.358208] env[61991]: _type = "Task" [ 634.358208] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.367321] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c92c93-9652-dadd-23bb-6a60fa2bdc2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.546832] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.645602] env[61991]: DEBUG nova.compute.manager [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Received event network-changed-b0a036d2-bd1d-4daa-9b89-cd85e219c8b5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.645814] env[61991]: DEBUG nova.compute.manager [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Refreshing instance network info cache due to event network-changed-b0a036d2-bd1d-4daa-9b89-cd85e219c8b5. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.646145] env[61991]: DEBUG oslo_concurrency.lockutils [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] Acquiring lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.646252] env[61991]: DEBUG oslo_concurrency.lockutils [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] Acquired lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.646342] env[61991]: DEBUG nova.network.neutron [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Refreshing network info cache for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 634.795189] env[61991]: DEBUG oslo_vmware.api [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129320, 'name': PowerOffVM_Task, 'duration_secs': 0.196897} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.795189] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 634.795189] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 634.795189] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-31f7d634-0aad-425a-bff1-3f667094a5ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.809626] env[61991]: ERROR nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 634.809626] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.809626] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.809626] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.809626] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.809626] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.809626] env[61991]: ERROR nova.compute.manager raise self.value [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.809626] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.809626] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.809626] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.810234] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.810234] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.810234] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 634.810234] env[61991]: ERROR nova.compute.manager [ 634.810234] env[61991]: Traceback (most recent call last): [ 634.810234] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.810234] env[61991]: listener.cb(fileno) [ 634.810234] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.810234] env[61991]: result = function(*args, **kwargs) [ 634.810234] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.810234] env[61991]: return func(*args, **kwargs) [ 634.810234] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.810234] env[61991]: raise e [ 634.810234] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.810234] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 634.810234] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.810234] env[61991]: created_port_ids = self._update_ports_for_instance( [ 634.810234] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.810234] env[61991]: with excutils.save_and_reraise_exception(): [ 634.810234] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.810234] env[61991]: self.force_reraise() [ 634.810234] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.810234] env[61991]: raise self.value [ 634.810234] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.810234] env[61991]: updated_port = self._update_port( [ 634.810234] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.810234] env[61991]: _ensure_no_port_binding_failure(port) [ 634.810234] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.810234] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.811102] env[61991]: nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 634.811102] env[61991]: Removing descriptor: 18 [ 634.818216] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 634.818443] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 634.818609] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Deleting the datastore file [datastore1] 432ab035-b5b0-4390-bef2-ec2691b4422b {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 634.818854] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31f35bc4-96ab-43cd-ab5f-17bfa5eccbe0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.824612] env[61991]: DEBUG oslo_vmware.api [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for the task: (returnval){ [ 634.824612] env[61991]: value = "task-1129322" [ 634.824612] env[61991]: _type = "Task" [ 634.824612] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.834141] env[61991]: DEBUG oslo_vmware.api [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.837734] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.838365] env[61991]: ERROR nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Traceback (most recent call last): [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self.driver.spawn(context, instance, image_meta, [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] vm_ref = self.build_virtual_machine(instance, [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.838365] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] for vif in network_info: [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] return self._sync_wrapper(fn, *args, **kwargs) [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self.wait() [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self[:] = self._gt.wait() [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] return self._exit_event.wait() [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] current.throw(*self._exc) [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.838712] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] result = function(*args, **kwargs) [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] return func(*args, **kwargs) [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] raise e [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] nwinfo = self.network_api.allocate_for_instance( [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] created_port_ids = self._update_ports_for_instance( [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] with excutils.save_and_reraise_exception(): [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] self.force_reraise() [ 634.839093] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] raise self.value [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] updated_port = self._update_port( [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] _ensure_no_port_binding_failure(port) [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] raise exception.PortBindingFailed(port_id=port['id']) [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] nova.exception.PortBindingFailed: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. [ 634.839415] env[61991]: ERROR nova.compute.manager [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] [ 634.839415] env[61991]: DEBUG nova.compute.utils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.840506] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.666s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.842188] env[61991]: INFO nova.compute.claims [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.847565] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Build of instance 3f9a4042-d9e4-4201-8386-18637c39795f was re-scheduled: Binding failed for port 6ea9f583-188d-4df3-b9f6-c2a97b3208ac, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.848070] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.848330] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.848511] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.848704] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.868668] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c92c93-9652-dadd-23bb-6a60fa2bdc2b, 'name': SearchDatastore_Task, 'duration_secs': 0.011448} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.869480] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b17d2f68-a2bb-4b22-bc78-5edc58b4928c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.874909] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 634.874909] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527f2f85-d4a9-7615-8d2c-2d23a2eacf9c" [ 634.874909] env[61991]: _type = "Task" [ 634.874909] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.882538] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527f2f85-d4a9-7615-8d2c-2d23a2eacf9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.887167] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 634.914044] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 634.914503] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 634.914772] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.915085] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 634.915351] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.915608] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 634.915917] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 634.916193] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 634.916472] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 634.916797] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 634.917096] env[61991]: DEBUG nova.virt.hardware [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.918209] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843b6ce7-4e10-4b23-a4cc-66bee7e1d591 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.926723] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702100f3-55e0-4447-9c8a-65905e49271e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.941986] env[61991]: ERROR nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Traceback (most recent call last): [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] yield resources [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self.driver.spawn(context, instance, image_meta, [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] vm_ref = self.build_virtual_machine(instance, [ 634.941986] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] for vif in network_info: [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] return self._sync_wrapper(fn, *args, **kwargs) [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self.wait() [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self[:] = self._gt.wait() [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] return self._exit_event.wait() [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.942365] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] current.throw(*self._exc) [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] result = function(*args, **kwargs) [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] return func(*args, **kwargs) [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] raise e [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] nwinfo = self.network_api.allocate_for_instance( [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] created_port_ids = self._update_ports_for_instance( [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] with excutils.save_and_reraise_exception(): [ 634.942704] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self.force_reraise() [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] raise self.value [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] updated_port = self._update_port( [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] _ensure_no_port_binding_failure(port) [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] raise exception.PortBindingFailed(port_id=port['id']) [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 634.943037] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] [ 634.943037] env[61991]: INFO nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Terminating instance [ 634.944438] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.070305] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.164599] env[61991]: DEBUG nova.network.neutron [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.251280] env[61991]: DEBUG nova.network.neutron [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.336689] env[61991]: DEBUG oslo_vmware.api [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Task: {'id': task-1129322, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101551} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.336951] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 635.337148] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 635.337447] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 635.337683] env[61991]: INFO nova.compute.manager [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Took 1.07 seconds to destroy the instance on the hypervisor. [ 635.337940] env[61991]: DEBUG oslo.service.loopingcall [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.338148] env[61991]: DEBUG nova.compute.manager [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.338245] env[61991]: DEBUG nova.network.neutron [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.353580] env[61991]: DEBUG nova.network.neutron [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.368806] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.385332] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527f2f85-d4a9-7615-8d2c-2d23a2eacf9c, 'name': SearchDatastore_Task, 'duration_secs': 0.019455} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.385596] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.385985] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 635.386102] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80694f41-9a5d-4127-b9b6-76cfba5d9deb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.392372] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 635.392372] env[61991]: value = "task-1129323" [ 635.392372] env[61991]: _type = "Task" [ 635.392372] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.399534] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.453831] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.754075] env[61991]: DEBUG oslo_concurrency.lockutils [req-a0333b6c-f36d-43e3-95a7-2a25e2407d85 req-a099203d-1fcf-48f9-a3ec-9427dcf6fd1d service nova] Releasing lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.754572] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquired lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.754786] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.856787] env[61991]: DEBUG nova.network.neutron [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.905480] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478655} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.906028] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.906266] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 635.906553] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83024b73-ec4a-4863-9a07-c295c1a15592 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.913463] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 635.913463] env[61991]: value = "task-1129324" [ 635.913463] env[61991]: _type = "Task" [ 635.913463] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.926457] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129324, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.957138] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Releasing lock "refresh_cache-3f9a4042-d9e4-4201-8386-18637c39795f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.957401] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 635.957578] env[61991]: DEBUG nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.957753] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.979434] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.210548] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6861f118-7105-4302-be35-0f3686841679 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.218324] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b79f936-0542-4d43-8926-954ebee0ac84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.249631] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61eff2b9-b74a-46cb-8c45-52cccce72e6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.257396] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973bec29-3ddd-4797-bc0d-d8c2846d703e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.273327] env[61991]: DEBUG nova.compute.provider_tree [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.277705] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.360399] env[61991]: INFO nova.compute.manager [-] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Took 1.02 seconds to deallocate network for instance. [ 636.369976] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.423160] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129324, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068596} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.423462] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 636.424276] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5531b703-95df-4087-8e68-3f66e8281aa8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.443383] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 636.443744] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bfb6e44-aeb5-4add-97df-463cb88b89e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.462367] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 636.462367] env[61991]: value = "task-1129325" [ 636.462367] env[61991]: _type = "Task" [ 636.462367] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.470092] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.482062] env[61991]: DEBUG nova.network.neutron [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.667591] env[61991]: DEBUG nova.compute.manager [req-173d791f-3efe-4340-a938-b8d8e24101f2 req-b9e7e796-f886-4a36-8aa8-0f0b9fe7e3a5 service nova] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Received event network-vif-deleted-b0a036d2-bd1d-4daa-9b89-cd85e219c8b5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.780479] env[61991]: DEBUG nova.scheduler.client.report [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.868196] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.872251] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Releasing lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.873310] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.873310] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.873310] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f1b8edd-7ed2-46db-a403-5508373efb35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.883072] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a1c5ed-8c9d-4f71-96f6-39c791964acd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.915619] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 878aee87-a1fc-4c0d-b385-44a98f790536 could not be found. [ 636.915896] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.916293] env[61991]: INFO nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Took 0.04 seconds to destroy the instance on the hypervisor. [ 636.916607] env[61991]: DEBUG oslo.service.loopingcall [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.916838] env[61991]: DEBUG nova.compute.manager [-] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.916933] env[61991]: DEBUG nova.network.neutron [-] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.936150] env[61991]: DEBUG nova.network.neutron [-] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.973327] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129325, 'name': ReconfigVM_Task, 'duration_secs': 0.277602} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.973760] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Reconfigured VM instance instance-0000001a to attach disk [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719/c2ef2717-4eac-428b-92a8-12a37d7c0719.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 636.975104] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28fbe7f1-adfe-4d96-8745-7c81237b56f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.982432] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 636.982432] env[61991]: value = "task-1129326" [ 636.982432] env[61991]: _type = "Task" [ 636.982432] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.989391] env[61991]: INFO nova.compute.manager [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3f9a4042-d9e4-4201-8386-18637c39795f] Took 1.03 seconds to deallocate network for instance. [ 636.998530] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129326, 'name': Rename_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.286369] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.289016] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.291225] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.495s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.441610] env[61991]: DEBUG nova.network.neutron [-] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.497315] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129326, 'name': Rename_Task, 'duration_secs': 0.137856} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.498021] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 637.498101] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e559970-2f4c-46c7-bb91-919cbffb9930 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.508502] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Waiting for the task: (returnval){ [ 637.508502] env[61991]: value = "task-1129327" [ 637.508502] env[61991]: _type = "Task" [ 637.508502] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.518769] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.797929] env[61991]: DEBUG nova.compute.utils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.804978] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 637.805117] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 637.858694] env[61991]: DEBUG nova.policy [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '90b9dc0e940f448f92f1807cf8d3ae29', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9074d948fbbb42e2909170a6e08d1ab2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.945246] env[61991]: INFO nova.compute.manager [-] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Took 1.03 seconds to deallocate network for instance. [ 637.947687] env[61991]: DEBUG nova.compute.claims [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.947867] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.019514] env[61991]: DEBUG oslo_vmware.api [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Task: {'id': task-1129327, 'name': PowerOnVM_Task, 'duration_secs': 0.405535} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.019855] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 638.020082] env[61991]: DEBUG nova.compute.manager [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 638.020877] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356a2e36-a0b3-4419-a3b0-98695938280d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.032149] env[61991]: INFO nova.scheduler.client.report [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Deleted allocations for instance 3f9a4042-d9e4-4201-8386-18637c39795f [ 638.212788] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Successfully created port: 2830fb24-0975-431c-ad82-ff4bc9b98b01 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.300981] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.335742] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c2ef2717-4eac-428b-92a8-12a37d7c0719 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.335742] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 432ab035-b5b0-4390-bef2-ec2691b4422b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 638.335888] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 32736198-a819-4afc-929b-a2c890263059 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.337218] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f720a778-5f29-4f27-8b84-7bf377f27dc9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.337632] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 931d58eb-3bbc-4d81-b737-95830b342027 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.337632] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 878aee87-a1fc-4c0d-b385-44a98f790536 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.337738] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 84b50fb2-7338-4da7-a1cb-d638c421a36d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.539517] env[61991]: DEBUG oslo_concurrency.lockutils [None req-00ae6601-cdfe-4469-8fa9-0a354b13607d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3f9a4042-d9e4-4201-8386-18637c39795f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.847s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.545082] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.843387] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 2866aec9-56fd-4e77-b651-9d85783fb94f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.043974] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.315027] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.340017] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.340017] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.340017] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.340470] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.340751] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.344019] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.344019] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.344019] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.344019] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.344019] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.344336] env[61991]: DEBUG nova.virt.hardware [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.344336] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e7b87d-741b-46f8-90d6-8bd9d9ce954a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.346698] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 598c0bd7-baf5-4d77-8d06-0f83a08b9685 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.353571] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc335d87-4993-4a3b-874e-3ae685d2eda9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.417583] env[61991]: DEBUG nova.compute.manager [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Received event network-changed-2830fb24-0975-431c-ad82-ff4bc9b98b01 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.417583] env[61991]: DEBUG nova.compute.manager [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Refreshing instance network info cache due to event network-changed-2830fb24-0975-431c-ad82-ff4bc9b98b01. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.417583] env[61991]: DEBUG oslo_concurrency.lockutils [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] Acquiring lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.417583] env[61991]: DEBUG oslo_concurrency.lockutils [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] Acquired lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.418333] env[61991]: DEBUG nova.network.neutron [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Refreshing network info cache for port 2830fb24-0975-431c-ad82-ff4bc9b98b01 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 639.537072] env[61991]: ERROR nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 639.537072] env[61991]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.537072] env[61991]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.537072] env[61991]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.537072] env[61991]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.537072] env[61991]: ERROR nova.compute.manager self.force_reraise() [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.537072] env[61991]: ERROR nova.compute.manager raise self.value [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.537072] env[61991]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.537072] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.537072] env[61991]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.537604] env[61991]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.537604] env[61991]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.537604] env[61991]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 639.537604] env[61991]: ERROR nova.compute.manager [ 639.537604] env[61991]: Traceback (most recent call last): [ 639.537604] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.537604] env[61991]: listener.cb(fileno) [ 639.537604] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.537604] env[61991]: result = function(*args, **kwargs) [ 639.537604] env[61991]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.537604] env[61991]: return func(*args, **kwargs) [ 639.537604] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.537604] env[61991]: raise e [ 639.537604] env[61991]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.537604] env[61991]: nwinfo = self.network_api.allocate_for_instance( [ 639.537604] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.537604] env[61991]: created_port_ids = self._update_ports_for_instance( [ 639.537604] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.537604] env[61991]: with excutils.save_and_reraise_exception(): [ 639.537604] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.537604] env[61991]: self.force_reraise() [ 639.537604] env[61991]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.537604] env[61991]: raise self.value [ 639.537604] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.537604] env[61991]: updated_port = self._update_port( [ 639.537604] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.537604] env[61991]: _ensure_no_port_binding_failure(port) [ 639.537604] env[61991]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.537604] env[61991]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.538427] env[61991]: nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 639.538427] env[61991]: Removing descriptor: 17 [ 639.538427] env[61991]: ERROR nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Traceback (most recent call last): [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] yield resources [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self.driver.spawn(context, instance, image_meta, [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.538427] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] vm_ref = self.build_virtual_machine(instance, [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] for vif in network_info: [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return self._sync_wrapper(fn, *args, **kwargs) [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self.wait() [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self[:] = self._gt.wait() [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return self._exit_event.wait() [ 639.538775] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] result = hub.switch() [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return self.greenlet.switch() [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] result = function(*args, **kwargs) [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return func(*args, **kwargs) [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] raise e [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] nwinfo = self.network_api.allocate_for_instance( [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 639.539160] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] created_port_ids = self._update_ports_for_instance( [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] with excutils.save_and_reraise_exception(): [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self.force_reraise() [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] raise self.value [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] updated_port = self._update_port( [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] _ensure_no_port_binding_failure(port) [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.539522] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] raise exception.PortBindingFailed(port_id=port['id']) [ 639.540061] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 639.540061] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] [ 639.540061] env[61991]: INFO nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Terminating instance [ 639.540513] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquiring lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.579420] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.849561] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance b1c1c4e4-a82b-4066-8ee8-342d5ad18d29 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.941112] env[61991]: DEBUG nova.network.neutron [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.983753] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "c2ef2717-4eac-428b-92a8-12a37d7c0719" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.983958] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "c2ef2717-4eac-428b-92a8-12a37d7c0719" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.984181] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "c2ef2717-4eac-428b-92a8-12a37d7c0719-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.984365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "c2ef2717-4eac-428b-92a8-12a37d7c0719-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.984530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "c2ef2717-4eac-428b-92a8-12a37d7c0719-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.987355] env[61991]: INFO nova.compute.manager [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Terminating instance [ 639.989066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "refresh_cache-c2ef2717-4eac-428b-92a8-12a37d7c0719" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.989219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquired lock "refresh_cache-c2ef2717-4eac-428b-92a8-12a37d7c0719" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.989380] env[61991]: DEBUG nova.network.neutron [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.066445] env[61991]: DEBUG nova.network.neutron [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.321976] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.322233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.356589] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance a2606541-ff39-4d03-bd19-d0ee8e035c08 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.506676] env[61991]: DEBUG nova.network.neutron [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.555472] env[61991]: DEBUG nova.network.neutron [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.569264] env[61991]: DEBUG oslo_concurrency.lockutils [req-90c3bfb8-28f1-419e-a2b8-6ba054a8e35a req-fc3819b8-1aa0-495c-99d8-edce7ee99337 service nova] Releasing lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.570210] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquired lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.570508] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.860098] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9d88002b-d5f0-4669-9a34-bfbf8f9169d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.059342] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Releasing lock "refresh_cache-c2ef2717-4eac-428b-92a8-12a37d7c0719" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.059782] env[61991]: DEBUG nova.compute.manager [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.059973] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.060870] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e4da02-4ff7-4bc6-b3eb-73b276c7813a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.068992] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 641.069245] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a2d565d-44df-482f-9480-9ed2e58d1c36 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.075884] env[61991]: DEBUG oslo_vmware.api [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 641.075884] env[61991]: value = "task-1129328" [ 641.075884] env[61991]: _type = "Task" [ 641.075884] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.084165] env[61991]: DEBUG oslo_vmware.api [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.090891] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.179487] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.362529] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 812d3491-0598-4171-b905-5304ebf6ae37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.505248] env[61991]: DEBUG nova.compute.manager [req-cac906ab-c846-48f8-bd86-180f8ece2e94 req-2694d0f7-4eda-4f83-b258-7862384e06e4 service nova] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Received event network-vif-deleted-2830fb24-0975-431c-ad82-ff4bc9b98b01 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.585992] env[61991]: DEBUG oslo_vmware.api [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129328, 'name': PowerOffVM_Task, 'duration_secs': 0.208363} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.586368] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 641.586425] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 641.586657] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2bdcf8a-b278-431e-a72c-85e86cc89eb7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.614791] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 641.614961] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 641.615157] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleting the datastore file [datastore2] c2ef2717-4eac-428b-92a8-12a37d7c0719 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 641.615423] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-285e6123-8108-45c0-9a0f-a07eefbcbc4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.621540] env[61991]: DEBUG oslo_vmware.api [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for the task: (returnval){ [ 641.621540] env[61991]: value = "task-1129330" [ 641.621540] env[61991]: _type = "Task" [ 641.621540] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.628960] env[61991]: DEBUG oslo_vmware.api [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.682597] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Releasing lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.682993] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.683209] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.683638] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0b6467e-9089-4fb6-845c-6e0484a7a5ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.692526] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ad2e4d-4ca0-41fb-9509-33b87a8b5cf0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.713580] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 84b50fb2-7338-4da7-a1cb-d638c421a36d could not be found. [ 641.713797] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.713976] env[61991]: INFO nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 641.714234] env[61991]: DEBUG oslo.service.loopingcall [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.714445] env[61991]: DEBUG nova.compute.manager [-] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.714539] env[61991]: DEBUG nova.network.neutron [-] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 641.735271] env[61991]: DEBUG nova.network.neutron [-] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.865379] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance acef131a-4263-42ad-964e-bb1e0cb21eec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.131677] env[61991]: DEBUG oslo_vmware.api [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Task: {'id': task-1129330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103547} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.131955] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 642.132147] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 642.132302] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.132475] env[61991]: INFO nova.compute.manager [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Took 1.07 seconds to destroy the instance on the hypervisor. [ 642.132707] env[61991]: DEBUG oslo.service.loopingcall [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.132891] env[61991]: DEBUG nova.compute.manager [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.132982] env[61991]: DEBUG nova.network.neutron [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.150013] env[61991]: DEBUG nova.network.neutron [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.237425] env[61991]: DEBUG nova.network.neutron [-] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.371179] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance fe78a8de-6174-47b3-9271-171c4ba76068 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.653029] env[61991]: DEBUG nova.network.neutron [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.712477] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "3e06e424-b64e-4fd2-8013-27760200c41d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.712477] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.740335] env[61991]: INFO nova.compute.manager [-] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Took 1.03 seconds to deallocate network for instance. [ 642.742688] env[61991]: DEBUG nova.compute.claims [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Aborting claim: {{(pid=61991) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.742688] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.874282] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.155235] env[61991]: INFO nova.compute.manager [-] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Took 1.02 seconds to deallocate network for instance. [ 643.377113] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c14d99a3-950d-44d0-b330-3031a1f8a2be has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.661997] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.880139] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 527561af-ba19-447f-a4ee-618824b9ce0f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.384904] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 43cb61e7-1748-40d9-a287-1179c8219c2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.888955] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 1715711a-72c5-4e86-88e0-cecf922cb42f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.392871] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.895706] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 0631634f-5c58-4ab0-b0dd-71821caf2fc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.398668] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f78ef63b-453e-45d3-959b-4b0c1922b53e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.901141] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 30bd843a-9d86-4de1-b011-600e2a93bcb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.404055] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3bcde671-5702-4b8a-8881-88eb7dfd0556 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.907404] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 80d05278-b516-4408-94b0-11bc93500b5c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.411680] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 78631e83-e574-4a90-b062-176b2d12d681 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.411862] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 648.412015] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 648.729808] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b457f5-6e12-4bcc-b00b-803d1a58f588 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.737520] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03c950a-5e33-4f25-ba65-5b296dd8f5bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.766130] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d339d0af-91dd-4ca1-a455-419478eb2486 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.773407] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2159f2-672d-42c5-a18d-0d79ded272a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.786955] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.290022] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.795058] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 649.795258] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.505s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.795528] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.270s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.797127] env[61991]: INFO nova.compute.claims [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.119414] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7fe2bb-56c2-4fee-bfa8-95ab86ea63d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.127777] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd48229-fda4-490a-ba84-b066db524f45 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.158033] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0e7964-ea31-424e-b7dc-1787bcfa0f15 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.165871] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1c1d03-98ed-490b-845e-db86b0e8c125 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.180028] env[61991]: DEBUG nova.compute.provider_tree [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.682154] env[61991]: DEBUG nova.scheduler.client.report [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.190641] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.191220] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 652.198100] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.160s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.199547] env[61991]: INFO nova.compute.claims [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.707612] env[61991]: DEBUG nova.compute.utils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.711366] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.711366] env[61991]: DEBUG nova.network.neutron [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 652.771764] env[61991]: DEBUG nova.policy [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4c0c8d586704325b3f63ec4d1416b14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd60902f6921d49298cf74ec32ce00145', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.105148] env[61991]: DEBUG nova.network.neutron [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Successfully created port: 93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.220017] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 653.560999] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a76339-1342-40de-aa72-d136e279fd30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.567989] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7c18ca-ab8d-4243-87f3-d6f8ba7e19dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.598376] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc513cf-576c-4fde-bc07-79462c360eb5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.606291] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc1566d-f9c7-4c0b-90f7-9426261c0ef7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.621188] env[61991]: DEBUG nova.compute.provider_tree [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.124912] env[61991]: DEBUG nova.scheduler.client.report [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.226941] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 654.251466] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 654.251732] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 654.251886] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.252079] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 654.252224] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.252368] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 654.252667] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 654.252861] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 654.253041] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 654.253203] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 654.253433] env[61991]: DEBUG nova.virt.hardware [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 654.254303] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dc1f15-714e-4fa3-ac26-f2d182fb24de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.262791] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d45ce95-89ad-46b9-b1b0-90e37040866e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.584611] env[61991]: DEBUG nova.compute.manager [req-dd064cd0-50ad-4fc5-bf4b-87d924e04464 req-024a5150-85db-4ec1-8c7c-ea0982a4dd11 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Received event network-vif-plugged-93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.584848] env[61991]: DEBUG oslo_concurrency.lockutils [req-dd064cd0-50ad-4fc5-bf4b-87d924e04464 req-024a5150-85db-4ec1-8c7c-ea0982a4dd11 service nova] Acquiring lock "2866aec9-56fd-4e77-b651-9d85783fb94f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.585120] env[61991]: DEBUG oslo_concurrency.lockutils [req-dd064cd0-50ad-4fc5-bf4b-87d924e04464 req-024a5150-85db-4ec1-8c7c-ea0982a4dd11 service nova] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.585225] env[61991]: DEBUG oslo_concurrency.lockutils [req-dd064cd0-50ad-4fc5-bf4b-87d924e04464 req-024a5150-85db-4ec1-8c7c-ea0982a4dd11 service nova] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.585390] env[61991]: DEBUG nova.compute.manager [req-dd064cd0-50ad-4fc5-bf4b-87d924e04464 req-024a5150-85db-4ec1-8c7c-ea0982a4dd11 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] No waiting events found dispatching network-vif-plugged-93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 654.585547] env[61991]: WARNING nova.compute.manager [req-dd064cd0-50ad-4fc5-bf4b-87d924e04464 req-024a5150-85db-4ec1-8c7c-ea0982a4dd11 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Received unexpected event network-vif-plugged-93128ac1-77ad-44e6-ab82-16754a4ec201 for instance with vm_state building and task_state spawning. [ 654.629441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.629980] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.632514] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.551s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.687401] env[61991]: DEBUG nova.network.neutron [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Successfully updated port: 93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 655.141549] env[61991]: DEBUG nova.compute.utils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.146461] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.146817] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 655.184222] env[61991]: DEBUG nova.policy [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c54258ac7b3e493d808c496531ca4d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fd3b8e07dce44a8b03543c3c0ca1bb5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.191525] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "refresh_cache-2866aec9-56fd-4e77-b651-9d85783fb94f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.191525] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquired lock "refresh_cache-2866aec9-56fd-4e77-b651-9d85783fb94f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.191525] env[61991]: DEBUG nova.network.neutron [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 655.533766] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4734c331-2367-43e8-886e-b604edeb3f5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.541333] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98786aa2-9df3-4cfd-9228-39067f3640b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.546395] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Successfully created port: 2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.573197] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0939f69-8c1b-4519-9b4b-b1947ad1bed7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.581371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7f8795-eb0e-4ead-8540-026aa06abca9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.593213] env[61991]: DEBUG nova.compute.provider_tree [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.647349] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.738086] env[61991]: DEBUG nova.network.neutron [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.096854] env[61991]: DEBUG nova.scheduler.client.report [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.160245] env[61991]: DEBUG nova.network.neutron [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Updating instance_info_cache with network_info: [{"id": "93128ac1-77ad-44e6-ab82-16754a4ec201", "address": "fa:16:3e:c3:aa:00", "network": {"id": "d5183351-9132-4404-b6d5-aa6fdc63c182", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-879157672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d60902f6921d49298cf74ec32ce00145", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93128ac1-77", "ovs_interfaceid": "93128ac1-77ad-44e6-ab82-16754a4ec201", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.602520] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.603305] env[61991]: ERROR nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] Traceback (most recent call last): [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self.driver.spawn(context, instance, image_meta, [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] vm_ref = self.build_virtual_machine(instance, [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.603305] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] for vif in network_info: [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] return self._sync_wrapper(fn, *args, **kwargs) [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self.wait() [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self[:] = self._gt.wait() [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] return self._exit_event.wait() [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] current.throw(*self._exc) [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.603735] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] result = function(*args, **kwargs) [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] return func(*args, **kwargs) [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] raise e [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] nwinfo = self.network_api.allocate_for_instance( [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] created_port_ids = self._update_ports_for_instance( [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] with excutils.save_and_reraise_exception(): [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] self.force_reraise() [ 656.604106] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] raise self.value [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] updated_port = self._update_port( [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] _ensure_no_port_binding_failure(port) [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] raise exception.PortBindingFailed(port_id=port['id']) [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] nova.exception.PortBindingFailed: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. [ 656.604475] env[61991]: ERROR nova.compute.manager [instance: 32736198-a819-4afc-929b-a2c890263059] [ 656.604475] env[61991]: DEBUG nova.compute.utils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.605855] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.598s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.610819] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Build of instance 32736198-a819-4afc-929b-a2c890263059 was re-scheduled: Binding failed for port fb9df8fe-9922-4c59-a08c-693af5bba9ed, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.611386] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.611659] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquiring lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.611828] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Acquired lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.612015] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.613901] env[61991]: DEBUG nova.compute.manager [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Received event network-changed-93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.614091] env[61991]: DEBUG nova.compute.manager [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Refreshing instance network info cache due to event network-changed-93128ac1-77ad-44e6-ab82-16754a4ec201. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.614272] env[61991]: DEBUG oslo_concurrency.lockutils [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] Acquiring lock "refresh_cache-2866aec9-56fd-4e77-b651-9d85783fb94f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.656485] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.662127] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Releasing lock "refresh_cache-2866aec9-56fd-4e77-b651-9d85783fb94f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.662415] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Instance network_info: |[{"id": "93128ac1-77ad-44e6-ab82-16754a4ec201", "address": "fa:16:3e:c3:aa:00", "network": {"id": "d5183351-9132-4404-b6d5-aa6fdc63c182", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-879157672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d60902f6921d49298cf74ec32ce00145", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93128ac1-77", "ovs_interfaceid": "93128ac1-77ad-44e6-ab82-16754a4ec201", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 656.662676] env[61991]: DEBUG oslo_concurrency.lockutils [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] Acquired lock "refresh_cache-2866aec9-56fd-4e77-b651-9d85783fb94f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.662845] env[61991]: DEBUG nova.network.neutron [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Refreshing network info cache for port 93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.664266] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:aa:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ebf80ce9-9885-46ed-ac23-310a98789a95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '93128ac1-77ad-44e6-ab82-16754a4ec201', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 656.672634] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Creating folder: Project (d60902f6921d49298cf74ec32ce00145). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 656.675451] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-949a2b11-aef9-46b0-a74e-541785ebe3d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.683760] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.684012] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.684183] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.684367] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.684670] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.684836] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.685061] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.685227] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.685394] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.685573] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.685785] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.686768] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247d7f6c-ea70-4d69-b213-c5c258af9d67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.690597] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Created folder: Project (d60902f6921d49298cf74ec32ce00145) in parent group-v246753. [ 656.690794] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Creating folder: Instances. Parent ref: group-v246770. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 656.691337] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-827ff694-8667-4863-bb11-be46586f4d3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.696536] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d9e711-a128-4dcf-8377-af745313bfa5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.712938] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Created folder: Instances in parent group-v246770. [ 656.712938] env[61991]: DEBUG oslo.service.loopingcall [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.712938] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 656.712938] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-599e9ce8-0748-4cf2-b3ba-d3b2d5c1bf47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.733992] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 656.733992] env[61991]: value = "task-1129333" [ 656.733992] env[61991]: _type = "Task" [ 656.733992] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.741520] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129333, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.129638] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.244048] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129333, 'name': CreateVM_Task, 'duration_secs': 0.338144} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.245934] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 657.249203] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.255598] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.255753] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.256078] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 657.256324] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03c1f349-9910-49eb-8a6c-2a2eded91465 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.264273] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 657.264273] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea5231-a85b-4278-8fa8-c31248e6103b" [ 657.264273] env[61991]: _type = "Task" [ 657.264273] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.276717] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea5231-a85b-4278-8fa8-c31248e6103b, 'name': SearchDatastore_Task, 'duration_secs': 0.010945} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.276991] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.277260] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 657.277546] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.277633] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.277808] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 657.278041] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1b0f461-66f7-4029-81ea-36ef1f6b4801 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.285535] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 657.286487] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 657.286487] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aef6bc2d-009a-4d40-9ead-16e3a57561e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.291661] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 657.291661] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52787726-3792-a738-00fe-eb1896e28f37" [ 657.291661] env[61991]: _type = "Task" [ 657.291661] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.301521] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52787726-3792-a738-00fe-eb1896e28f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.521823] env[61991]: DEBUG nova.network.neutron [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Updated VIF entry in instance network info cache for port 93128ac1-77ad-44e6-ab82-16754a4ec201. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 657.522258] env[61991]: DEBUG nova.network.neutron [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Updating instance_info_cache with network_info: [{"id": "93128ac1-77ad-44e6-ab82-16754a4ec201", "address": "fa:16:3e:c3:aa:00", "network": {"id": "d5183351-9132-4404-b6d5-aa6fdc63c182", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-879157672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d60902f6921d49298cf74ec32ce00145", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ebf80ce9-9885-46ed-ac23-310a98789a95", "external-id": "nsx-vlan-transportzone-582", "segmentation_id": 582, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93128ac1-77", "ovs_interfaceid": "93128ac1-77ad-44e6-ab82-16754a4ec201", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.552721] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2785aa53-6159-4c8e-8d70-499c4b127060 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.559723] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7bdb19-0a3e-446c-8b9f-ddb7871c11dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.598292] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Successfully updated port: 2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 657.599966] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199abeec-c1f5-4273-9732-954b5fb6c9bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.608456] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af804ae-b328-4a46-af4c-5843effcad3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.622625] env[61991]: DEBUG nova.compute.provider_tree [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.753386] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Releasing lock "refresh_cache-32736198-a819-4afc-929b-a2c890263059" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.753703] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.753898] env[61991]: DEBUG nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.754076] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.768827] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.802425] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52787726-3792-a738-00fe-eb1896e28f37, 'name': SearchDatastore_Task, 'duration_secs': 0.00841} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.803233] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4169ae7a-981c-4083-bd42-344db43e3206 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.808083] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 657.808083] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524fa15b-6a45-d46a-f046-d1a030dcffca" [ 657.808083] env[61991]: _type = "Task" [ 657.808083] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.815853] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524fa15b-6a45-d46a-f046-d1a030dcffca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.025711] env[61991]: DEBUG oslo_concurrency.lockutils [req-8e9a4da6-c88a-4375-aaf0-9ac73b6e4674 req-a1bc9025-6894-40b2-b60f-66f1de87b8f9 service nova] Releasing lock "refresh_cache-2866aec9-56fd-4e77-b651-9d85783fb94f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.104571] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "refresh_cache-598c0bd7-baf5-4d77-8d06-0f83a08b9685" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.104776] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "refresh_cache-598c0bd7-baf5-4d77-8d06-0f83a08b9685" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.104949] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.125540] env[61991]: DEBUG nova.scheduler.client.report [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.270824] env[61991]: DEBUG nova.network.neutron [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.319132] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524fa15b-6a45-d46a-f046-d1a030dcffca, 'name': SearchDatastore_Task, 'duration_secs': 0.008856} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.319409] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.319658] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 2866aec9-56fd-4e77-b651-9d85783fb94f/2866aec9-56fd-4e77-b651-9d85783fb94f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 658.319914] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50668445-8579-4014-9a71-901d8c1086da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.328240] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 658.328240] env[61991]: value = "task-1129334" [ 658.328240] env[61991]: _type = "Task" [ 658.328240] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.336518] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129334, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.629968] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.024s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.630692] env[61991]: ERROR nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Traceback (most recent call last): [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self.driver.spawn(context, instance, image_meta, [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] vm_ref = self.build_virtual_machine(instance, [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.630692] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] for vif in network_info: [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] return self._sync_wrapper(fn, *args, **kwargs) [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self.wait() [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self[:] = self._gt.wait() [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] return self._exit_event.wait() [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] current.throw(*self._exc) [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.631050] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] result = function(*args, **kwargs) [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] return func(*args, **kwargs) [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] raise e [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] nwinfo = self.network_api.allocate_for_instance( [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] created_port_ids = self._update_ports_for_instance( [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] with excutils.save_and_reraise_exception(): [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] self.force_reraise() [ 658.631427] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] raise self.value [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] updated_port = self._update_port( [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] _ensure_no_port_binding_failure(port) [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] raise exception.PortBindingFailed(port_id=port['id']) [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] nova.exception.PortBindingFailed: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. [ 658.631812] env[61991]: ERROR nova.compute.manager [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] [ 658.631812] env[61991]: DEBUG nova.compute.utils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.632763] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.929s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.632963] env[61991]: DEBUG nova.objects.instance [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 658.636200] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Build of instance f720a778-5f29-4f27-8b84-7bf377f27dc9 was re-scheduled: Binding failed for port 9d03dcc3-8f67-4c34-921f-d236df62870d, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 658.636777] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 658.637128] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.637331] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquired lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.637591] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.657037] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.672996] env[61991]: DEBUG nova.compute.manager [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Received event network-vif-plugged-2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.673235] env[61991]: DEBUG oslo_concurrency.lockutils [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] Acquiring lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.673460] env[61991]: DEBUG oslo_concurrency.lockutils [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.673697] env[61991]: DEBUG oslo_concurrency.lockutils [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.673837] env[61991]: DEBUG nova.compute.manager [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] No waiting events found dispatching network-vif-plugged-2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 658.673991] env[61991]: WARNING nova.compute.manager [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Received unexpected event network-vif-plugged-2348155c-3fd0-4166-b513-8cc0e36278cf for instance with vm_state building and task_state spawning. [ 658.674205] env[61991]: DEBUG nova.compute.manager [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Received event network-changed-2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 658.674318] env[61991]: DEBUG nova.compute.manager [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Refreshing instance network info cache due to event network-changed-2348155c-3fd0-4166-b513-8cc0e36278cf. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 658.674925] env[61991]: DEBUG oslo_concurrency.lockutils [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] Acquiring lock "refresh_cache-598c0bd7-baf5-4d77-8d06-0f83a08b9685" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.775737] env[61991]: INFO nova.compute.manager [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] [instance: 32736198-a819-4afc-929b-a2c890263059] Took 1.02 seconds to deallocate network for instance. [ 658.837756] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Updating instance_info_cache with network_info: [{"id": "2348155c-3fd0-4166-b513-8cc0e36278cf", "address": "fa:16:3e:bd:c3:2e", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348155c-3f", "ovs_interfaceid": "2348155c-3fd0-4166-b513-8cc0e36278cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.846505] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129334, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468227} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.846758] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 2866aec9-56fd-4e77-b651-9d85783fb94f/2866aec9-56fd-4e77-b651-9d85783fb94f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 658.846963] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 658.847218] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a513aea8-1cd0-413a-b53e-d9da7b77ad8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.855122] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 658.855122] env[61991]: value = "task-1129335" [ 658.855122] env[61991]: _type = "Task" [ 658.855122] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.864205] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129335, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.159721] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.237144] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.339382] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "refresh_cache-598c0bd7-baf5-4d77-8d06-0f83a08b9685" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.339964] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Instance network_info: |[{"id": "2348155c-3fd0-4166-b513-8cc0e36278cf", "address": "fa:16:3e:bd:c3:2e", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348155c-3f", "ovs_interfaceid": "2348155c-3fd0-4166-b513-8cc0e36278cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 659.340111] env[61991]: DEBUG oslo_concurrency.lockutils [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] Acquired lock "refresh_cache-598c0bd7-baf5-4d77-8d06-0f83a08b9685" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.340213] env[61991]: DEBUG nova.network.neutron [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Refreshing network info cache for port 2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 659.342107] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:c3:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2348155c-3fd0-4166-b513-8cc0e36278cf', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 659.348911] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Creating folder: Project (5fd3b8e07dce44a8b03543c3c0ca1bb5). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.349892] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-169c906a-1f03-4a31-a111-1c5d057df994 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.361089] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Created folder: Project (5fd3b8e07dce44a8b03543c3c0ca1bb5) in parent group-v246753. [ 659.361284] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Creating folder: Instances. Parent ref: group-v246773. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.364347] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47d9fb6a-607a-41a5-b07b-2ddd10b9d756 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.365919] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129335, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066196} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.366176] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 659.367216] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320e2086-261b-44d7-a905-afa0320a6801 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.389784] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 2866aec9-56fd-4e77-b651-9d85783fb94f/2866aec9-56fd-4e77-b651-9d85783fb94f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 659.390199] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7ce5899-f152-43a1-8306-4ea632cafcd7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.403946] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Created folder: Instances in parent group-v246773. [ 659.404200] env[61991]: DEBUG oslo.service.loopingcall [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.404399] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 659.404970] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34001679-5307-46fd-85df-e7851fc1bcf8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.420023] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 659.420023] env[61991]: value = "task-1129338" [ 659.420023] env[61991]: _type = "Task" [ 659.420023] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.424653] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 659.424653] env[61991]: value = "task-1129339" [ 659.424653] env[61991]: _type = "Task" [ 659.424653] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.430840] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129338, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.435258] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129339, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.648372] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31a528df-4660-4286-b3e7-d69d94fe9a14 tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.650066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.191s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.651213] env[61991]: INFO nova.compute.claims [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.739940] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Releasing lock "refresh_cache-f720a778-5f29-4f27-8b84-7bf377f27dc9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.739940] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 659.739940] env[61991]: DEBUG nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.739940] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.754937] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.806755] env[61991]: INFO nova.scheduler.client.report [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Deleted allocations for instance 32736198-a819-4afc-929b-a2c890263059 [ 659.934168] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129338, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.937152] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129339, 'name': CreateVM_Task, 'duration_secs': 0.36284} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.937296] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 659.937907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.938080] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.938392] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 659.938629] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44ceeefa-9335-4e15-8a5d-779589a64bb0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.944792] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 659.944792] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52de6455-5c1d-f3b7-9546-34c126494f89" [ 659.944792] env[61991]: _type = "Task" [ 659.944792] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.952327] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52de6455-5c1d-f3b7-9546-34c126494f89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.096031] env[61991]: DEBUG nova.network.neutron [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Updated VIF entry in instance network info cache for port 2348155c-3fd0-4166-b513-8cc0e36278cf. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 660.096031] env[61991]: DEBUG nova.network.neutron [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Updating instance_info_cache with network_info: [{"id": "2348155c-3fd0-4166-b513-8cc0e36278cf", "address": "fa:16:3e:bd:c3:2e", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2348155c-3f", "ovs_interfaceid": "2348155c-3fd0-4166-b513-8cc0e36278cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.257561] env[61991]: DEBUG nova.network.neutron [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.318709] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eb02dfda-4fe7-4387-886b-8d4627cf03c1 tempest-ServersTestBootFromVolume-559589408 tempest-ServersTestBootFromVolume-559589408-project-member] Lock "32736198-a819-4afc-929b-a2c890263059" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.062s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.430645] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129338, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.455118] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52de6455-5c1d-f3b7-9546-34c126494f89, 'name': SearchDatastore_Task, 'duration_secs': 0.010655} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.455390] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.455619] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 660.455846] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.455991] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.456183] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 660.456432] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed731ad7-74fe-4d43-8146-4f03bb1c3249 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.464539] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 660.464796] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 660.465486] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80180e1d-23dd-474f-bed3-4ebf50bd2c49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.470163] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 660.470163] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c1f196-fe74-f195-06b3-9a456f9110d8" [ 660.470163] env[61991]: _type = "Task" [ 660.470163] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.477770] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c1f196-fe74-f195-06b3-9a456f9110d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.598260] env[61991]: DEBUG oslo_concurrency.lockutils [req-d040bf21-5793-4095-b2a2-1601611b5df1 req-7d6d434c-7696-468c-beef-c4763e035652 service nova] Releasing lock "refresh_cache-598c0bd7-baf5-4d77-8d06-0f83a08b9685" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.759670] env[61991]: INFO nova.compute.manager [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: f720a778-5f29-4f27-8b84-7bf377f27dc9] Took 1.02 seconds to deallocate network for instance. [ 660.823449] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.934034] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129338, 'name': ReconfigVM_Task, 'duration_secs': 1.294586} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.934175] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 2866aec9-56fd-4e77-b651-9d85783fb94f/2866aec9-56fd-4e77-b651-9d85783fb94f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 660.934823] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6449e55e-5ad0-498d-aa2e-981c1ac74fcf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.941156] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 660.941156] env[61991]: value = "task-1129340" [ 660.941156] env[61991]: _type = "Task" [ 660.941156] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.953850] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129340, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.982295] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c1f196-fe74-f195-06b3-9a456f9110d8, 'name': SearchDatastore_Task, 'duration_secs': 0.00852} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.985705] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fadde263-fc4c-408e-a4ad-1f7ab046af92 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.992585] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 660.992585] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284d984-20b9-e242-6176-9ab3308a9874" [ 660.992585] env[61991]: _type = "Task" [ 660.992585] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.003439] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284d984-20b9-e242-6176-9ab3308a9874, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.008514] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163b4706-0771-4f1f-b5a4-ac2f69f01c0b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.015178] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1112daf-991c-4b07-9fe9-23881a9dab67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.045999] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068b3950-8ac9-48e3-9292-b02319594562 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.053154] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17f28ca-360b-4332-924f-20258ca68f77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.065985] env[61991]: DEBUG nova.compute.provider_tree [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.342902] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.452399] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129340, 'name': Rename_Task, 'duration_secs': 0.138499} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.453030] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 661.453406] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70b080ca-eecb-418d-adb7-e95564a4305f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.461514] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 661.461514] env[61991]: value = "task-1129342" [ 661.461514] env[61991]: _type = "Task" [ 661.461514] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.469552] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129342, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.502774] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284d984-20b9-e242-6176-9ab3308a9874, 'name': SearchDatastore_Task, 'duration_secs': 0.011685} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.503049] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.503307] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 598c0bd7-baf5-4d77-8d06-0f83a08b9685/598c0bd7-baf5-4d77-8d06-0f83a08b9685.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 661.503625] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a94290f6-2723-4932-9499-d0be6e44319b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.511466] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 661.511466] env[61991]: value = "task-1129343" [ 661.511466] env[61991]: _type = "Task" [ 661.511466] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.519445] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129343, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.569610] env[61991]: DEBUG nova.scheduler.client.report [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.793263] env[61991]: INFO nova.scheduler.client.report [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Deleted allocations for instance f720a778-5f29-4f27-8b84-7bf377f27dc9 [ 661.970690] env[61991]: DEBUG oslo_vmware.api [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129342, 'name': PowerOnVM_Task, 'duration_secs': 0.493636} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.971175] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 661.971175] env[61991]: INFO nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Took 7.74 seconds to spawn the instance on the hypervisor. [ 661.971350] env[61991]: DEBUG nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 661.972141] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eeefacb-c3f5-4be9-bfb2-b0ae9bb6d501 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.021567] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129343, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.077028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.077028] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.078108] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.712s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.078565] env[61991]: DEBUG nova.objects.instance [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 662.303826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c943404b-88f4-4936-a40b-5dc0486e8704 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "f720a778-5f29-4f27-8b84-7bf377f27dc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.257s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.493668] env[61991]: INFO nova.compute.manager [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Took 36.99 seconds to build instance. [ 662.527036] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129343, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.590026] env[61991]: DEBUG nova.compute.utils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.590026] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.590026] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.653933] env[61991]: DEBUG nova.policy [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c54258ac7b3e493d808c496531ca4d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fd3b8e07dce44a8b03543c3c0ca1bb5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.806273] env[61991]: DEBUG nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.995495] env[61991]: DEBUG oslo_concurrency.lockutils [None req-abd25129-0306-42d8-8153-34fff8f2638a tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.950s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.031576] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129343, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.116937} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.032559] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 598c0bd7-baf5-4d77-8d06-0f83a08b9685/598c0bd7-baf5-4d77-8d06-0f83a08b9685.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 663.032844] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 663.033431] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77e6321a-ec37-4306-924b-967627347e32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.042188] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 663.042188] env[61991]: value = "task-1129344" [ 663.042188] env[61991]: _type = "Task" [ 663.042188] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.048916] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.091749] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3eb30d83-8b16-4fb9-8496-7cfdc788f2f6 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.092883] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.407s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.097130] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.099986] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "2866aec9-56fd-4e77-b651-9d85783fb94f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.100278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.100653] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "2866aec9-56fd-4e77-b651-9d85783fb94f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.100850] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.101051] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.103298] env[61991]: INFO nova.compute.manager [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Terminating instance [ 663.110782] env[61991]: DEBUG nova.compute.manager [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.110782] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.110782] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45dc95a7-ef8e-4bb2-b03f-8094a8b27cb2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.116079] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 663.116322] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6d1bb76-3850-4f4e-ae5d-c7cbd4b4caff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.124319] env[61991]: DEBUG oslo_vmware.api [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 663.124319] env[61991]: value = "task-1129345" [ 663.124319] env[61991]: _type = "Task" [ 663.124319] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.133196] env[61991]: DEBUG oslo_vmware.api [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129345, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.205105] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Successfully created port: 30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.334271] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.498703] env[61991]: DEBUG nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.552280] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064776} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.552823] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 663.553342] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13c4413-a01c-45e6-bed6-0fd0fcc89136 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.579655] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] 598c0bd7-baf5-4d77-8d06-0f83a08b9685/598c0bd7-baf5-4d77-8d06-0f83a08b9685.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 663.580337] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3a899cd-e958-471e-9e32-c40c6c93612c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.612802] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 663.612802] env[61991]: value = "task-1129346" [ 663.612802] env[61991]: _type = "Task" [ 663.612802] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.623513] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129346, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.636370] env[61991]: DEBUG oslo_vmware.api [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129345, 'name': PowerOffVM_Task, 'duration_secs': 0.331968} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.636703] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 663.636919] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 663.637168] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16d2ab86-aec1-4fd1-a2ee-24ae20493e2d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.698912] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 663.698981] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 663.699583] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Deleting the datastore file [datastore2] 2866aec9-56fd-4e77-b651-9d85783fb94f {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 663.699583] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81c7ab73-c375-432f-9a47-82426f47cbfb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.711095] env[61991]: DEBUG oslo_vmware.api [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for the task: (returnval){ [ 663.711095] env[61991]: value = "task-1129348" [ 663.711095] env[61991]: _type = "Task" [ 663.711095] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.720172] env[61991]: DEBUG oslo_vmware.api [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.016659] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e3c9b8-a726-4cc0-911a-f28a8f363837 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.020044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.025241] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae77fbe-c3ba-46c3-b5e1-681924c276bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.056817] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce9685e-9637-4c4f-96bd-76d506f50608 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.064667] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be88a06-0f29-49e7-9726-179862251f8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.078722] env[61991]: DEBUG nova.compute.provider_tree [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.114677] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.128933] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129346, 'name': ReconfigVM_Task, 'duration_secs': 0.272441} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.129282] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Reconfigured VM instance instance-00000023 to attach disk [datastore2] 598c0bd7-baf5-4d77-8d06-0f83a08b9685/598c0bd7-baf5-4d77-8d06-0f83a08b9685.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 664.129946] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4d0b7c6-7c4c-4687-8085-c0c07231a99f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.137992] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 664.137992] env[61991]: value = "task-1129349" [ 664.137992] env[61991]: _type = "Task" [ 664.137992] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.147195] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129349, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.149557] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.149769] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.149916] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.150103] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.150241] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.150379] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.150576] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.150725] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.150881] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.151042] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.151208] env[61991]: DEBUG nova.virt.hardware [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.151990] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc44cc1-3bb6-4cd2-bb14-2d2ce3996dc5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.160286] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60987bd9-9bd8-4d94-a37d-04579ec96204 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.223060] env[61991]: DEBUG oslo_vmware.api [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Task: {'id': task-1129348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168813} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.224647] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 664.224833] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 664.225018] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.225180] env[61991]: INFO nova.compute.manager [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 664.225428] env[61991]: DEBUG oslo.service.loopingcall [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.228134] env[61991]: DEBUG nova.compute.manager [-] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.228252] env[61991]: DEBUG nova.network.neutron [-] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.231158] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.231279] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.531159] env[61991]: DEBUG nova.compute.manager [req-af91e216-1b06-423e-b98e-28ac6f07a914 req-edc1643f-2315-4d99-ba4d-5576dce671b5 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Received event network-vif-deleted-93128ac1-77ad-44e6-ab82-16754a4ec201 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.531388] env[61991]: INFO nova.compute.manager [req-af91e216-1b06-423e-b98e-28ac6f07a914 req-edc1643f-2315-4d99-ba4d-5576dce671b5 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Neutron deleted interface 93128ac1-77ad-44e6-ab82-16754a4ec201; detaching it from the instance and deleting it from the info cache [ 664.531687] env[61991]: DEBUG nova.network.neutron [req-af91e216-1b06-423e-b98e-28ac6f07a914 req-edc1643f-2315-4d99-ba4d-5576dce671b5 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.582333] env[61991]: DEBUG nova.scheduler.client.report [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.648112] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129349, 'name': Rename_Task, 'duration_secs': 0.138845} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.648393] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 664.648819] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3ecbc42-9f88-4e06-a4a7-8677ef9a290e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.656062] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 664.656062] env[61991]: value = "task-1129350" [ 664.656062] env[61991]: _type = "Task" [ 664.656062] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.664435] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.819133] env[61991]: DEBUG nova.compute.manager [req-e774720d-844d-468d-afb5-00beb9d9393e req-556246ae-6600-4f8e-a35d-802d11213383 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Received event network-vif-plugged-30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.819399] env[61991]: DEBUG oslo_concurrency.lockutils [req-e774720d-844d-468d-afb5-00beb9d9393e req-556246ae-6600-4f8e-a35d-802d11213383 service nova] Acquiring lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.819612] env[61991]: DEBUG oslo_concurrency.lockutils [req-e774720d-844d-468d-afb5-00beb9d9393e req-556246ae-6600-4f8e-a35d-802d11213383 service nova] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.819793] env[61991]: DEBUG oslo_concurrency.lockutils [req-e774720d-844d-468d-afb5-00beb9d9393e req-556246ae-6600-4f8e-a35d-802d11213383 service nova] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.819954] env[61991]: DEBUG nova.compute.manager [req-e774720d-844d-468d-afb5-00beb9d9393e req-556246ae-6600-4f8e-a35d-802d11213383 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] No waiting events found dispatching network-vif-plugged-30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 664.822106] env[61991]: WARNING nova.compute.manager [req-e774720d-844d-468d-afb5-00beb9d9393e req-556246ae-6600-4f8e-a35d-802d11213383 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Received unexpected event network-vif-plugged-30d912af-abee-4191-bff0-574811acc2f5 for instance with vm_state building and task_state spawning. [ 664.917140] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Successfully updated port: 30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 664.988191] env[61991]: DEBUG nova.network.neutron [-] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.035830] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9779a727-a7f5-45e1-bbf5-48ed31ed508c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.046923] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982756a6-66f7-46c3-bb27-c70495705a74 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.069355] env[61991]: DEBUG nova.compute.manager [req-af91e216-1b06-423e-b98e-28ac6f07a914 req-edc1643f-2315-4d99-ba4d-5576dce671b5 service nova] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Detach interface failed, port_id=93128ac1-77ad-44e6-ab82-16754a4ec201, reason: Instance 2866aec9-56fd-4e77-b651-9d85783fb94f could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 665.089945] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.090710] env[61991]: ERROR nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Traceback (most recent call last): [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self.driver.spawn(context, instance, image_meta, [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] vm_ref = self.build_virtual_machine(instance, [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.090710] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] for vif in network_info: [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return self._sync_wrapper(fn, *args, **kwargs) [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self.wait() [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self[:] = self._gt.wait() [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return self._exit_event.wait() [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] result = hub.switch() [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.091101] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return self.greenlet.switch() [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] result = function(*args, **kwargs) [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] return func(*args, **kwargs) [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] raise e [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] nwinfo = self.network_api.allocate_for_instance( [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] created_port_ids = self._update_ports_for_instance( [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] with excutils.save_and_reraise_exception(): [ 665.091592] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] self.force_reraise() [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] raise self.value [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] updated_port = self._update_port( [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] _ensure_no_port_binding_failure(port) [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] raise exception.PortBindingFailed(port_id=port['id']) [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] nova.exception.PortBindingFailed: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. [ 665.092046] env[61991]: ERROR nova.compute.manager [instance: 931d58eb-3bbc-4d81-b737-95830b342027] [ 665.092359] env[61991]: DEBUG nova.compute.utils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.092814] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.023s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.097985] env[61991]: INFO nova.compute.claims [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.099996] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Build of instance 931d58eb-3bbc-4d81-b737-95830b342027 was re-scheduled: Binding failed for port 409facee-38ec-442f-9b89-3afc939a8e18, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.100454] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.100673] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.100816] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquired lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.100973] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.167099] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129350, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.425295] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "refresh_cache-b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.425295] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "refresh_cache-b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.425295] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.492508] env[61991]: INFO nova.compute.manager [-] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Took 1.26 seconds to deallocate network for instance. [ 665.623780] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.666516] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129350, 'name': PowerOnVM_Task, 'duration_secs': 0.840053} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.666763] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 665.666986] env[61991]: INFO nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Took 9.01 seconds to spawn the instance on the hypervisor. [ 665.667180] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 665.667957] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9770cdb3-d19e-40a6-868f-56bcb5ed6ff4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.725787] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.957549] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.004426] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.122032] env[61991]: DEBUG nova.network.neutron [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Updating instance_info_cache with network_info: [{"id": "30d912af-abee-4191-bff0-574811acc2f5", "address": "fa:16:3e:12:36:bd", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d912af-ab", "ovs_interfaceid": "30d912af-abee-4191-bff0-574811acc2f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.191747] env[61991]: INFO nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Took 40.17 seconds to build instance. [ 666.228998] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Releasing lock "refresh_cache-931d58eb-3bbc-4d81-b737-95830b342027" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.229281] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.229973] env[61991]: DEBUG nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.230279] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.245779] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.502437] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ac699f-b65c-49fc-a6bf-d157bf4e695e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.510273] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08150bc6-36a1-425c-906d-63c4cbc7da38 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.541040] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8e4674-46ac-48e3-8aab-a29973b06623 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.548010] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1538cb3-c46b-499a-bc4a-34c32198475f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.564267] env[61991]: DEBUG nova.compute.provider_tree [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.625489] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "refresh_cache-b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.625831] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Instance network_info: |[{"id": "30d912af-abee-4191-bff0-574811acc2f5", "address": "fa:16:3e:12:36:bd", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d912af-ab", "ovs_interfaceid": "30d912af-abee-4191-bff0-574811acc2f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 666.626261] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:36:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30d912af-abee-4191-bff0-574811acc2f5', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.633589] env[61991]: DEBUG oslo.service.loopingcall [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.633800] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 666.633800] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-231135be-f771-4a12-924d-0e3ada73f4ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.655259] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.655259] env[61991]: value = "task-1129351" [ 666.655259] env[61991]: _type = "Task" [ 666.655259] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.664340] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129351, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.694127] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.901s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.750314] env[61991]: DEBUG nova.network.neutron [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.866747] env[61991]: DEBUG nova.compute.manager [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Received event network-changed-30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.867388] env[61991]: DEBUG nova.compute.manager [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Refreshing instance network info cache due to event network-changed-30d912af-abee-4191-bff0-574811acc2f5. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 666.867388] env[61991]: DEBUG oslo_concurrency.lockutils [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] Acquiring lock "refresh_cache-b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.867535] env[61991]: DEBUG oslo_concurrency.lockutils [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] Acquired lock "refresh_cache-b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.867891] env[61991]: DEBUG nova.network.neutron [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Refreshing network info cache for port 30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 667.069049] env[61991]: DEBUG nova.scheduler.client.report [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.166846] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129351, 'name': CreateVM_Task, 'duration_secs': 0.287143} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.167453] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 667.167743] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.167910] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.168244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 667.168495] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ebe7917-039b-433a-9f48-4594414435c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.173266] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 667.173266] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a25dba-0f9b-2500-37af-a532507464ae" [ 667.173266] env[61991]: _type = "Task" [ 667.173266] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.181621] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a25dba-0f9b-2500-37af-a532507464ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.197031] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.253396] env[61991]: INFO nova.compute.manager [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 931d58eb-3bbc-4d81-b737-95830b342027] Took 1.02 seconds to deallocate network for instance. [ 667.575766] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.576398] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.579402] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.712s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.581511] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.581511] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.634s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.627494] env[61991]: DEBUG nova.network.neutron [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Updated VIF entry in instance network info cache for port 30d912af-abee-4191-bff0-574811acc2f5. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 667.627899] env[61991]: DEBUG nova.network.neutron [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Updating instance_info_cache with network_info: [{"id": "30d912af-abee-4191-bff0-574811acc2f5", "address": "fa:16:3e:12:36:bd", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30d912af-ab", "ovs_interfaceid": "30d912af-abee-4191-bff0-574811acc2f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.633112] env[61991]: INFO nova.scheduler.client.report [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Deleted allocations for instance 432ab035-b5b0-4390-bef2-ec2691b4422b [ 667.684093] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a25dba-0f9b-2500-37af-a532507464ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009628} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.684372] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.684620] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 667.684863] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.685018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.685200] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 667.685449] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0fde43a-0cd9-41b3-8b90-081e96aeead3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.693609] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 667.694555] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 667.694555] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2337721c-f315-48f8-b5be-a15793ba221e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.699786] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 667.699786] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521d150f-572f-1093-bb1d-07aa8505f79a" [ 667.699786] env[61991]: _type = "Task" [ 667.699786] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.713557] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521d150f-572f-1093-bb1d-07aa8505f79a, 'name': SearchDatastore_Task, 'duration_secs': 0.00924} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.715586] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0087c43-19bd-4dd9-b8e1-c235374a2457 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.719379] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 667.719379] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea954e-a235-c4f3-64a8-a6cc14923083" [ 667.719379] env[61991]: _type = "Task" [ 667.719379] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.723255] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.727951] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea954e-a235-c4f3-64a8-a6cc14923083, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.085972] env[61991]: DEBUG nova.compute.utils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.096074] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.096565] env[61991]: DEBUG nova.network.neutron [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 668.103272] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.130657] env[61991]: DEBUG oslo_concurrency.lockutils [req-1dfee66c-6533-4e51-b8b5-12303de3be04 req-f00ac05a-db34-4336-9a2e-75fbaef7f945 service nova] Releasing lock "refresh_cache-b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.146957] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d692394-7a31-4af0-8ab7-4279ca257bb3 tempest-ServerShowV254Test-1425573419 tempest-ServerShowV254Test-1425573419-project-member] Lock "432ab035-b5b0-4390-bef2-ec2691b4422b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.003s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.164173] env[61991]: DEBUG nova.policy [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '55af68f4e91f4d50ae84048713f76036', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b19371ee81046b2931d054717635332', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.238260] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea954e-a235-c4f3-64a8-a6cc14923083, 'name': SearchDatastore_Task, 'duration_secs': 0.008667} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.238260] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.238260] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b1c1c4e4-a82b-4066-8ee8-342d5ad18d29/b1c1c4e4-a82b-4066-8ee8-342d5ad18d29.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 668.238260] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5657384-17b0-47f6-823f-c84a56b0a811 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.245415] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 668.245415] env[61991]: value = "task-1129352" [ 668.245415] env[61991]: _type = "Task" [ 668.245415] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.258264] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129352, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.284486] env[61991]: INFO nova.scheduler.client.report [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Deleted allocations for instance 931d58eb-3bbc-4d81-b737-95830b342027 [ 668.551554] env[61991]: DEBUG nova.network.neutron [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Successfully created port: 4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.578794] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52dae0cf-dbb1-4fed-9973-2a4fac9816bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.590027] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4076271b-abe2-4c26-a47e-80f762feeb23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.626618] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0237e72-6c5d-4655-8253-c0c36590bd6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.639789] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587ebbdf-66dc-498a-a4ee-40c0f4496738 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.657174] env[61991]: DEBUG nova.compute.provider_tree [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.755218] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129352, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455442} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.755567] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b1c1c4e4-a82b-4066-8ee8-342d5ad18d29/b1c1c4e4-a82b-4066-8ee8-342d5ad18d29.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 668.755693] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 668.755947] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd5c2fdf-8717-42aa-b7d5-525b332e5698 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.763738] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 668.763738] env[61991]: value = "task-1129353" [ 668.763738] env[61991]: _type = "Task" [ 668.763738] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.773675] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129353, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.797120] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc64897e-bb6d-4678-9c56-a356d02d2ea4 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "931d58eb-3bbc-4d81-b737-95830b342027" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.469s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.132795] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.161463] env[61991]: DEBUG nova.scheduler.client.report [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.169154] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.169154] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.169154] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.169523] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.169523] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.169523] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.169785] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.169990] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.170212] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.170412] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.170692] env[61991]: DEBUG nova.virt.hardware [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.171533] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6bef04-94a9-4ea5-a504-03cba1865e31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.181569] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab81d85-00a8-4204-9bca-16afb7d72dec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.273800] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129353, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060237} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.274097] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.275695] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d762fea-0263-4948-aa3c-d967bbc71eae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.299082] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] b1c1c4e4-a82b-4066-8ee8-342d5ad18d29/b1c1c4e4-a82b-4066-8ee8-342d5ad18d29.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.299756] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.302058] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-131ececa-bb00-47b7-8986-d506d4af8f84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.322646] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 669.322646] env[61991]: value = "task-1129354" [ 669.322646] env[61991]: _type = "Task" [ 669.322646] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.331993] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129354, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.677025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.094s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.677025] env[61991]: ERROR nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Traceback (most recent call last): [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self.driver.spawn(context, instance, image_meta, [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.677025] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] vm_ref = self.build_virtual_machine(instance, [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] for vif in network_info: [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] return self._sync_wrapper(fn, *args, **kwargs) [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self.wait() [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self[:] = self._gt.wait() [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] return self._exit_event.wait() [ 669.677456] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] current.throw(*self._exc) [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] result = function(*args, **kwargs) [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] return func(*args, **kwargs) [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] raise e [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] nwinfo = self.network_api.allocate_for_instance( [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] created_port_ids = self._update_ports_for_instance( [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.677891] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] with excutils.save_and_reraise_exception(): [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] self.force_reraise() [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] raise self.value [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] updated_port = self._update_port( [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] _ensure_no_port_binding_failure(port) [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] raise exception.PortBindingFailed(port_id=port['id']) [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] nova.exception.PortBindingFailed: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. [ 669.678457] env[61991]: ERROR nova.compute.manager [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] [ 669.678803] env[61991]: DEBUG nova.compute.utils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.680372] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.135s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.680372] env[61991]: DEBUG nova.objects.instance [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 669.683463] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Build of instance 878aee87-a1fc-4c0d-b385-44a98f790536 was re-scheduled: Binding failed for port b0a036d2-bd1d-4daa-9b89-cd85e219c8b5, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 669.683967] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 669.684255] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquiring lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.684453] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Acquired lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.684685] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.834639] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129354, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.839177] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.208918] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.242537] env[61991]: DEBUG nova.compute.manager [req-f9151d28-03ac-4a42-80b2-8751a69d7ecb req-2eebdba8-85ce-40a6-8a3c-35571e5af768 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Received event network-vif-plugged-4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.242741] env[61991]: DEBUG oslo_concurrency.lockutils [req-f9151d28-03ac-4a42-80b2-8751a69d7ecb req-2eebdba8-85ce-40a6-8a3c-35571e5af768 service nova] Acquiring lock "a2606541-ff39-4d03-bd19-d0ee8e035c08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.242938] env[61991]: DEBUG oslo_concurrency.lockutils [req-f9151d28-03ac-4a42-80b2-8751a69d7ecb req-2eebdba8-85ce-40a6-8a3c-35571e5af768 service nova] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.244474] env[61991]: DEBUG oslo_concurrency.lockutils [req-f9151d28-03ac-4a42-80b2-8751a69d7ecb req-2eebdba8-85ce-40a6-8a3c-35571e5af768 service nova] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.244688] env[61991]: DEBUG nova.compute.manager [req-f9151d28-03ac-4a42-80b2-8751a69d7ecb req-2eebdba8-85ce-40a6-8a3c-35571e5af768 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] No waiting events found dispatching network-vif-plugged-4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 670.244859] env[61991]: WARNING nova.compute.manager [req-f9151d28-03ac-4a42-80b2-8751a69d7ecb req-2eebdba8-85ce-40a6-8a3c-35571e5af768 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Received unexpected event network-vif-plugged-4b335179-f726-4551-a93c-01b6472885e1 for instance with vm_state building and task_state spawning. [ 670.311832] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.335783] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129354, 'name': ReconfigVM_Task, 'duration_secs': 0.738937} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.336146] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Reconfigured VM instance instance-00000024 to attach disk [datastore1] b1c1c4e4-a82b-4066-8ee8-342d5ad18d29/b1c1c4e4-a82b-4066-8ee8-342d5ad18d29.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 670.337233] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1299d6ea-1a85-4b7c-a170-d15829173f5b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.346017] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 670.346017] env[61991]: value = "task-1129355" [ 670.346017] env[61991]: _type = "Task" [ 670.346017] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.352687] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129355, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.388989] env[61991]: DEBUG nova.network.neutron [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Successfully updated port: 4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 670.563136] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "7df78da9-bf65-4621-b50d-43f1d721c2f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.563365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.696702] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adc20758-7e0d-47b3-9f74-4130257915f8 tempest-ServersAdmin275Test-551238688 tempest-ServersAdmin275Test-551238688-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.697914] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.119s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.703241] env[61991]: INFO nova.compute.claims [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.815764] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Releasing lock "refresh_cache-878aee87-a1fc-4c0d-b385-44a98f790536" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.816192] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.816495] env[61991]: DEBUG nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.816823] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.831287] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.854410] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129355, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.894220] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "refresh_cache-a2606541-ff39-4d03-bd19-d0ee8e035c08" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.894369] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquired lock "refresh_cache-a2606541-ff39-4d03-bd19-d0ee8e035c08" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.894522] env[61991]: DEBUG nova.network.neutron [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.335256] env[61991]: DEBUG nova.network.neutron [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.358405] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129355, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.427698] env[61991]: DEBUG nova.network.neutron [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.601747] env[61991]: DEBUG nova.network.neutron [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Updating instance_info_cache with network_info: [{"id": "4b335179-f726-4551-a93c-01b6472885e1", "address": "fa:16:3e:03:3d:85", "network": {"id": "c366bd6f-49ca-430a-a8c3-f322e290cb1d", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1975768760-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b19371ee81046b2931d054717635332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b335179-f7", "ovs_interfaceid": "4b335179-f726-4551-a93c-01b6472885e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.837167] env[61991]: INFO nova.compute.manager [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] [instance: 878aee87-a1fc-4c0d-b385-44a98f790536] Took 1.02 seconds to deallocate network for instance. [ 671.855561] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129355, 'name': Rename_Task, 'duration_secs': 1.137777} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.857844] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 671.858257] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff04b765-2275-4e23-87e8-72b959f2e290 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.864712] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 671.864712] env[61991]: value = "task-1129356" [ 671.864712] env[61991]: _type = "Task" [ 671.864712] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.872779] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.029595] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa10158-8080-4549-aa33-3bff9322da69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.037078] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ffc6ba-c7ed-49c6-8c50-d1ca5bcd746b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.068029] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891f3740-0647-4874-b45e-f8e557123a07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.077503] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7035ff-c201-493b-9680-80489d4d5d47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.091816] env[61991]: DEBUG nova.compute.provider_tree [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.108425] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Releasing lock "refresh_cache-a2606541-ff39-4d03-bd19-d0ee8e035c08" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.108741] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Instance network_info: |[{"id": "4b335179-f726-4551-a93c-01b6472885e1", "address": "fa:16:3e:03:3d:85", "network": {"id": "c366bd6f-49ca-430a-a8c3-f322e290cb1d", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1975768760-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b19371ee81046b2931d054717635332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b335179-f7", "ovs_interfaceid": "4b335179-f726-4551-a93c-01b6472885e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 672.109585] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:3d:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74f30339-6421-4654-bddb-81d7f34db9d7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b335179-f726-4551-a93c-01b6472885e1', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 672.116659] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Creating folder: Project (1b19371ee81046b2931d054717635332). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 672.116923] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f16c9ce-b0a2-442e-bdaa-f3abaf32aa8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.127426] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Created folder: Project (1b19371ee81046b2931d054717635332) in parent group-v246753. [ 672.127672] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Creating folder: Instances. Parent ref: group-v246777. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 672.127894] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa147d8a-4efc-4254-8181-23eca8f4f59b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.138688] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Created folder: Instances in parent group-v246777. [ 672.138919] env[61991]: DEBUG oslo.service.loopingcall [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.139113] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 672.139338] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0b2e333-e9ba-4972-9920-48cc5a903815 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.158396] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 672.158396] env[61991]: value = "task-1129359" [ 672.158396] env[61991]: _type = "Task" [ 672.158396] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.165957] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129359, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.270611] env[61991]: DEBUG nova.compute.manager [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Received event network-changed-4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.270932] env[61991]: DEBUG nova.compute.manager [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Refreshing instance network info cache due to event network-changed-4b335179-f726-4551-a93c-01b6472885e1. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 672.271051] env[61991]: DEBUG oslo_concurrency.lockutils [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] Acquiring lock "refresh_cache-a2606541-ff39-4d03-bd19-d0ee8e035c08" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.271225] env[61991]: DEBUG oslo_concurrency.lockutils [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] Acquired lock "refresh_cache-a2606541-ff39-4d03-bd19-d0ee8e035c08" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.271334] env[61991]: DEBUG nova.network.neutron [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Refreshing network info cache for port 4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 672.374836] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129356, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.595184] env[61991]: DEBUG nova.scheduler.client.report [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.668425] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129359, 'name': CreateVM_Task, 'duration_secs': 0.295584} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.668607] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 672.670088] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.670088] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.670088] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 672.670088] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0110c29c-5e23-4d3a-8ff7-5a6961ae5d6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.675270] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 672.675270] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520bde60-f408-2063-62e3-bad22067bd8b" [ 672.675270] env[61991]: _type = "Task" [ 672.675270] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.686700] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520bde60-f408-2063-62e3-bad22067bd8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.865794] env[61991]: INFO nova.scheduler.client.report [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Deleted allocations for instance 878aee87-a1fc-4c0d-b385-44a98f790536 [ 672.883388] env[61991]: DEBUG oslo_vmware.api [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129356, 'name': PowerOnVM_Task, 'duration_secs': 0.850838} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.883671] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 672.883883] env[61991]: INFO nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Took 8.77 seconds to spawn the instance on the hypervisor. [ 672.884066] env[61991]: DEBUG nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 672.884869] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96f872a-1b8d-4054-b7b8-99470b16c731 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.024135] env[61991]: DEBUG nova.network.neutron [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Updated VIF entry in instance network info cache for port 4b335179-f726-4551-a93c-01b6472885e1. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 673.024479] env[61991]: DEBUG nova.network.neutron [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Updating instance_info_cache with network_info: [{"id": "4b335179-f726-4551-a93c-01b6472885e1", "address": "fa:16:3e:03:3d:85", "network": {"id": "c366bd6f-49ca-430a-a8c3-f322e290cb1d", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1975768760-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1b19371ee81046b2931d054717635332", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74f30339-6421-4654-bddb-81d7f34db9d7", "external-id": "nsx-vlan-transportzone-899", "segmentation_id": 899, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b335179-f7", "ovs_interfaceid": "4b335179-f726-4551-a93c-01b6472885e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.099883] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.100516] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.107024] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.360s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.186026] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520bde60-f408-2063-62e3-bad22067bd8b, 'name': SearchDatastore_Task, 'duration_secs': 0.028265} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.186237] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.186469] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 673.186703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.186850] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.187031] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 673.187291] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9074549f-b2b3-4496-8483-a9c0061ebcbe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.195538] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 673.195722] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 673.196445] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb1ed784-36f5-47fc-b234-e824288cc3a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.201809] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 673.201809] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a8b2b-ee7e-248d-3ae1-c869d783f1f8" [ 673.201809] env[61991]: _type = "Task" [ 673.201809] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.210274] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a8b2b-ee7e-248d-3ae1-c869d783f1f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.384238] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27a37abf-8efb-43f2-b291-6a011a95e8e2 tempest-ServersAdminTestJSON-1263057964 tempest-ServersAdminTestJSON-1263057964-project-member] Lock "878aee87-a1fc-4c0d-b385-44a98f790536" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.175s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.400783] env[61991]: INFO nova.compute.manager [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Took 42.96 seconds to build instance. [ 673.527141] env[61991]: DEBUG oslo_concurrency.lockutils [req-97380ed3-7500-4fd6-afee-13fcd68e8145 req-760c2c7a-f210-48c6-bcbe-23e48fb2fae2 service nova] Releasing lock "refresh_cache-a2606541-ff39-4d03-bd19-d0ee8e035c08" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.608537] env[61991]: DEBUG nova.compute.utils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.614266] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.614266] env[61991]: DEBUG nova.network.neutron [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 673.661791] env[61991]: DEBUG nova.policy [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db4a19098fe7445b8552bdb7253e76d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '403c79445dbe46609a260dbbf92386b5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.711698] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a8b2b-ee7e-248d-3ae1-c869d783f1f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008828} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.714895] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a8a36ff-d6ba-41db-9b74-f5ea9606b6dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.720965] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 673.720965] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528beff2-45ea-69a0-6e23-f8d5f830cd64" [ 673.720965] env[61991]: _type = "Task" [ 673.720965] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.730777] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528beff2-45ea-69a0-6e23-f8d5f830cd64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.891098] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.904498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-85365c60-06ba-4e4b-867d-95fdadf44bf8 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.080s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.944255] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760e2eaf-e49d-496b-9784-386dec0d8ff3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.955821] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6075c29-e750-48e9-a242-49cb1581140e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.991077] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e65c74-bb42-45c5-950e-ee81848bcc6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.999835] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72b88d8-cd84-4727-aada-e5422df94b19 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.013086] env[61991]: DEBUG nova.compute.provider_tree [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.082703] env[61991]: DEBUG nova.network.neutron [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Successfully created port: ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.114131] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.231645] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528beff2-45ea-69a0-6e23-f8d5f830cd64, 'name': SearchDatastore_Task, 'duration_secs': 0.009196} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.234354] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.234354] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] a2606541-ff39-4d03-bd19-d0ee8e035c08/a2606541-ff39-4d03-bd19-d0ee8e035c08.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 674.234354] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-14a1c259-68b7-4dbf-88f9-6d5a1b3a5be0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.238938] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 674.238938] env[61991]: value = "task-1129360" [ 674.238938] env[61991]: _type = "Task" [ 674.238938] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.246947] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129360, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.409972] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.427850] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.516197] env[61991]: DEBUG nova.scheduler.client.report [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.750934] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129360, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.791736] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.792063] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.792283] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.792465] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.792635] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.796197] env[61991]: INFO nova.compute.manager [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Terminating instance [ 674.798582] env[61991]: DEBUG nova.compute.manager [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.798687] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.799558] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5788145e-efc6-42ce-9f2d-30940cfef373 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.807252] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 674.807771] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edbee171-04cb-409a-99ca-62a0770a1022 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.814152] env[61991]: DEBUG oslo_vmware.api [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 674.814152] env[61991]: value = "task-1129361" [ 674.814152] env[61991]: _type = "Task" [ 674.814152] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.822252] env[61991]: DEBUG oslo_vmware.api [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.945556] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.956313] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.956779] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.957037] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.957230] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.957379] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.959982] env[61991]: INFO nova.compute.manager [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Terminating instance [ 674.962150] env[61991]: DEBUG nova.compute.manager [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.962354] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.963230] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e8c95c-9d86-42c9-8433-86f17b8a78b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.973659] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 674.973659] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d258177-2562-496b-a709-bd14f8b39260 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.982572] env[61991]: DEBUG oslo_vmware.api [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 674.982572] env[61991]: value = "task-1129362" [ 674.982572] env[61991]: _type = "Task" [ 674.982572] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.994516] env[61991]: DEBUG oslo_vmware.api [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.021711] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.022416] env[61991]: ERROR nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Traceback (most recent call last): [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self.driver.spawn(context, instance, image_meta, [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] vm_ref = self.build_virtual_machine(instance, [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.022416] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] for vif in network_info: [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return self._sync_wrapper(fn, *args, **kwargs) [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self.wait() [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self[:] = self._gt.wait() [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return self._exit_event.wait() [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] result = hub.switch() [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.022779] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return self.greenlet.switch() [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] result = function(*args, **kwargs) [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] return func(*args, **kwargs) [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] raise e [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] nwinfo = self.network_api.allocate_for_instance( [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] created_port_ids = self._update_ports_for_instance( [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] with excutils.save_and_reraise_exception(): [ 675.023156] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] self.force_reraise() [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] raise self.value [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] updated_port = self._update_port( [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] _ensure_no_port_binding_failure(port) [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] raise exception.PortBindingFailed(port_id=port['id']) [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] nova.exception.PortBindingFailed: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. [ 675.023546] env[61991]: ERROR nova.compute.manager [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] [ 675.023876] env[61991]: DEBUG nova.compute.utils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.024882] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.363s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.025188] env[61991]: DEBUG nova.objects.instance [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lazy-loading 'resources' on Instance uuid c2ef2717-4eac-428b-92a8-12a37d7c0719 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 675.030475] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Build of instance 84b50fb2-7338-4da7-a1cb-d638c421a36d was re-scheduled: Binding failed for port 2830fb24-0975-431c-ad82-ff4bc9b98b01, please check neutron logs for more information. {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.032663] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Unplugging VIFs for instance {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.032973] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquiring lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.033139] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Acquired lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.033303] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.125237] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.159579] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.159949] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.160362] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.160654] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.160877] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.161108] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.161473] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.161708] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.161947] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.162186] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.162429] env[61991]: DEBUG nova.virt.hardware [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.163724] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33f79c4-d45e-4ada-ae9c-c704d02ea114 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.173276] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02211860-a59a-4438-a3be-a8da24c74d33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.249195] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129360, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.920752} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.249477] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] a2606541-ff39-4d03-bd19-d0ee8e035c08/a2606541-ff39-4d03-bd19-d0ee8e035c08.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 675.249676] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 675.249928] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f6d39c9-c617-43b1-9f89-122c8638bd38 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.257747] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 675.257747] env[61991]: value = "task-1129363" [ 675.257747] env[61991]: _type = "Task" [ 675.257747] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.265963] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129363, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.323966] env[61991]: DEBUG oslo_vmware.api [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129361, 'name': PowerOffVM_Task, 'duration_secs': 0.230046} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.324353] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 675.324514] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 675.324788] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf258110-7d90-4acb-94be-47e0aac7454a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.409958] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 675.410222] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 675.410478] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleting the datastore file [datastore2] 598c0bd7-baf5-4d77-8d06-0f83a08b9685 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.410710] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e55ae127-5bc3-4cf5-84a4-a3cc01219356 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.419184] env[61991]: DEBUG oslo_vmware.api [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 675.419184] env[61991]: value = "task-1129365" [ 675.419184] env[61991]: _type = "Task" [ 675.419184] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.433331] env[61991]: DEBUG oslo_vmware.api [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.491405] env[61991]: DEBUG oslo_vmware.api [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129362, 'name': PowerOffVM_Task, 'duration_secs': 0.304708} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.491858] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 675.492150] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 675.492441] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4ef3805-039a-4d17-9695-d5caf4576825 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.568319] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.676651] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.769801] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129363, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067821} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.773597] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 675.774615] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e133f5-39cf-4a70-a861-750fe84ec1eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.803879] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] a2606541-ff39-4d03-bd19-d0ee8e035c08/a2606541-ff39-4d03-bd19-d0ee8e035c08.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 675.806517] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd1c898a-fbd7-44e6-a87c-d777059def7d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.826564] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 675.826564] env[61991]: value = "task-1129367" [ 675.826564] env[61991]: _type = "Task" [ 675.826564] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.836331] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.877887] env[61991]: DEBUG nova.compute.manager [req-78b180f3-6741-491a-b7b7-b13984a48226 req-6d0a304d-965e-46c9-ba32-2c047502171e service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Received event network-vif-plugged-ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.878303] env[61991]: DEBUG oslo_concurrency.lockutils [req-78b180f3-6741-491a-b7b7-b13984a48226 req-6d0a304d-965e-46c9-ba32-2c047502171e service nova] Acquiring lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.878628] env[61991]: DEBUG oslo_concurrency.lockutils [req-78b180f3-6741-491a-b7b7-b13984a48226 req-6d0a304d-965e-46c9-ba32-2c047502171e service nova] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.878895] env[61991]: DEBUG oslo_concurrency.lockutils [req-78b180f3-6741-491a-b7b7-b13984a48226 req-6d0a304d-965e-46c9-ba32-2c047502171e service nova] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.879199] env[61991]: DEBUG nova.compute.manager [req-78b180f3-6741-491a-b7b7-b13984a48226 req-6d0a304d-965e-46c9-ba32-2c047502171e service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] No waiting events found dispatching network-vif-plugged-ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 675.879389] env[61991]: WARNING nova.compute.manager [req-78b180f3-6741-491a-b7b7-b13984a48226 req-6d0a304d-965e-46c9-ba32-2c047502171e service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Received unexpected event network-vif-plugged-ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 for instance with vm_state building and task_state spawning. [ 675.931860] env[61991]: DEBUG oslo_vmware.api [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145296} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.932133] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 675.932365] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 675.932501] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 675.932656] env[61991]: INFO nova.compute.manager [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Took 1.13 seconds to destroy the instance on the hypervisor. [ 675.932911] env[61991]: DEBUG oslo.service.loopingcall [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.933149] env[61991]: DEBUG nova.compute.manager [-] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.933301] env[61991]: DEBUG nova.network.neutron [-] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.962428] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e411075-9417-467c-b2c8-bddb997f8692 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.972980] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde52ca8-156b-46fe-8afb-7ee7c191e8ef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.009604] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252b391a-427b-4397-9db3-64ee2ef33b96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.018277] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ddeedf-f6bf-40b2-b7f9-aceeeb9e4798 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.035599] env[61991]: DEBUG nova.compute.provider_tree [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.038643] env[61991]: DEBUG nova.network.neutron [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Successfully updated port: ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 676.182148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Releasing lock "refresh_cache-84b50fb2-7338-4da7-a1cb-d638c421a36d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.182148] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61991) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.182148] env[61991]: DEBUG nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.182148] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.204186] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.230417] env[61991]: DEBUG nova.compute.manager [req-1ec3f36b-fd98-4679-99d2-666ecc3512a5 req-6b5abaa7-cfe1-46f6-bb33-e96588833ef2 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Received event network-vif-deleted-2348155c-3fd0-4166-b513-8cc0e36278cf {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.230508] env[61991]: INFO nova.compute.manager [req-1ec3f36b-fd98-4679-99d2-666ecc3512a5 req-6b5abaa7-cfe1-46f6-bb33-e96588833ef2 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Neutron deleted interface 2348155c-3fd0-4166-b513-8cc0e36278cf; detaching it from the instance and deleting it from the info cache [ 676.230659] env[61991]: DEBUG nova.network.neutron [req-1ec3f36b-fd98-4679-99d2-666ecc3512a5 req-6b5abaa7-cfe1-46f6-bb33-e96588833ef2 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.338852] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129367, 'name': ReconfigVM_Task, 'duration_secs': 0.29433} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.339199] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Reconfigured VM instance instance-00000025 to attach disk [datastore1] a2606541-ff39-4d03-bd19-d0ee8e035c08/a2606541-ff39-4d03-bd19-d0ee8e035c08.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 676.339869] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26a99dcb-f3f0-4a85-9155-e17af032b259 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.347817] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 676.347817] env[61991]: value = "task-1129368" [ 676.347817] env[61991]: _type = "Task" [ 676.347817] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.355915] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129368, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.541048] env[61991]: DEBUG nova.scheduler.client.report [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.544453] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.544614] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquired lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.544713] env[61991]: DEBUG nova.network.neutron [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.710588] env[61991]: DEBUG nova.network.neutron [-] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.711076] env[61991]: DEBUG nova.network.neutron [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.737793] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32dae7e9-8785-4e33-b75c-eb2ada3d7ce1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.748354] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ff4aae-00d5-495c-86f0-87ace4e1f9d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.760781] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 676.761321] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 676.761531] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleting the datastore file [datastore1] b1c1c4e4-a82b-4066-8ee8-342d5ad18d29 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 676.762406] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c44eb189-d9d9-442e-83d6-82919014f115 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.768228] env[61991]: DEBUG oslo_vmware.api [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 676.768228] env[61991]: value = "task-1129369" [ 676.768228] env[61991]: _type = "Task" [ 676.768228] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.774963] env[61991]: DEBUG nova.compute.manager [req-1ec3f36b-fd98-4679-99d2-666ecc3512a5 req-6b5abaa7-cfe1-46f6-bb33-e96588833ef2 service nova] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Detach interface failed, port_id=2348155c-3fd0-4166-b513-8cc0e36278cf, reason: Instance 598c0bd7-baf5-4d77-8d06-0f83a08b9685 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 676.779829] env[61991]: DEBUG oslo_vmware.api [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.857629] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129368, 'name': Rename_Task, 'duration_secs': 0.135503} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.858255] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 676.858506] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f05330b6-a49a-4d42-9152-757c3f4ca1f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.864916] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 676.864916] env[61991]: value = "task-1129370" [ 676.864916] env[61991]: _type = "Task" [ 676.864916] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.872382] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.048083] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.023s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.053028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.709s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.054816] env[61991]: INFO nova.compute.claims [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.087883] env[61991]: INFO nova.scheduler.client.report [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Deleted allocations for instance c2ef2717-4eac-428b-92a8-12a37d7c0719 [ 677.108706] env[61991]: DEBUG nova.network.neutron [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.212685] env[61991]: INFO nova.compute.manager [-] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Took 1.28 seconds to deallocate network for instance. [ 677.213180] env[61991]: INFO nova.compute.manager [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] [instance: 84b50fb2-7338-4da7-a1cb-d638c421a36d] Took 1.03 seconds to deallocate network for instance. [ 677.279546] env[61991]: DEBUG oslo_vmware.api [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138795} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.279824] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 677.279979] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 677.280179] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 677.280348] env[61991]: INFO nova.compute.manager [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Took 2.32 seconds to destroy the instance on the hypervisor. [ 677.280583] env[61991]: DEBUG oslo.service.loopingcall [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.280779] env[61991]: DEBUG nova.compute.manager [-] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.280876] env[61991]: DEBUG nova.network.neutron [-] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.374841] env[61991]: DEBUG oslo_vmware.api [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129370, 'name': PowerOnVM_Task, 'duration_secs': 0.460478} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.375626] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 677.375912] env[61991]: INFO nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Took 8.24 seconds to spawn the instance on the hypervisor. [ 677.376117] env[61991]: DEBUG nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 677.377025] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08828d1-acd7-4984-9abf-2f1402928fad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.421695] env[61991]: DEBUG nova.network.neutron [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updating instance_info_cache with network_info: [{"id": "ea0fc9ad-ea54-4ccb-bce4-3c95855ec049", "address": "fa:16:3e:08:84:d5", "network": {"id": "0fccf457-0b11-4a14-a61a-493e55641197", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1825044270-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "403c79445dbe46609a260dbbf92386b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea0fc9ad-ea", "ovs_interfaceid": "ea0fc9ad-ea54-4ccb-bce4-3c95855ec049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.599697] env[61991]: DEBUG oslo_concurrency.lockutils [None req-84bfee91-1fac-43f5-9f68-0b230156b8db tempest-ServersAdmin275Test-743385453 tempest-ServersAdmin275Test-743385453-project-member] Lock "c2ef2717-4eac-428b-92a8-12a37d7c0719" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.615s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.723157] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.899306] env[61991]: INFO nova.compute.manager [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Took 42.85 seconds to build instance. [ 677.923214] env[61991]: DEBUG nova.compute.manager [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Received event network-changed-ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.923411] env[61991]: DEBUG nova.compute.manager [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Refreshing instance network info cache due to event network-changed-ea0fc9ad-ea54-4ccb-bce4-3c95855ec049. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 677.923699] env[61991]: DEBUG oslo_concurrency.lockutils [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] Acquiring lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.924748] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Releasing lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.925067] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Instance network_info: |[{"id": "ea0fc9ad-ea54-4ccb-bce4-3c95855ec049", "address": "fa:16:3e:08:84:d5", "network": {"id": "0fccf457-0b11-4a14-a61a-493e55641197", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1825044270-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "403c79445dbe46609a260dbbf92386b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea0fc9ad-ea", "ovs_interfaceid": "ea0fc9ad-ea54-4ccb-bce4-3c95855ec049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 677.925343] env[61991]: DEBUG oslo_concurrency.lockutils [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] Acquired lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.925521] env[61991]: DEBUG nova.network.neutron [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Refreshing network info cache for port ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 677.928041] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:84:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea0fc9ad-ea54-4ccb-bce4-3c95855ec049', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 677.934280] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Creating folder: Project (403c79445dbe46609a260dbbf92386b5). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 677.935485] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f055990f-94d4-46b3-a24e-0e30386f3b43 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.947471] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Created folder: Project (403c79445dbe46609a260dbbf92386b5) in parent group-v246753. [ 677.947471] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Creating folder: Instances. Parent ref: group-v246780. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 677.947471] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d416e9e-106c-4b30-a4ab-7f1e35b8e788 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.956144] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Created folder: Instances in parent group-v246780. [ 677.956393] env[61991]: DEBUG oslo.service.loopingcall [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.956575] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 677.956771] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d154443-e71e-45f0-8b5c-d4b46f8bdffd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.977878] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 677.977878] env[61991]: value = "task-1129373" [ 677.977878] env[61991]: _type = "Task" [ 677.977878] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.985281] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129373, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.055980] env[61991]: DEBUG nova.network.neutron [-] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.262185] env[61991]: INFO nova.scheduler.client.report [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Deleted allocations for instance 84b50fb2-7338-4da7-a1cb-d638c421a36d [ 678.333210] env[61991]: DEBUG nova.compute.manager [req-2b025136-a7b0-40a8-960c-5c7c7ea83086 req-a81917ba-458c-47e5-b41e-29c882acbda1 service nova] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Received event network-vif-deleted-30d912af-abee-4191-bff0-574811acc2f5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.404174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e59e5991-b871-4924-8dff-cdeb57cfaec1 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.210s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.498282] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129373, 'name': CreateVM_Task, 'duration_secs': 0.313108} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.498796] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 678.499732] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.499929] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.500245] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 678.500487] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3bf2f3e-2379-492f-a46f-b0bf508b7adf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.505797] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 678.505797] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52279389-e0f5-af82-ec87-a97f0e14278d" [ 678.505797] env[61991]: _type = "Task" [ 678.505797] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.511609] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efee2b7c-ae17-408e-9e86-182e83df1b82 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.521374] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52279389-e0f5-af82-ec87-a97f0e14278d, 'name': SearchDatastore_Task, 'duration_secs': 0.009078} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.523385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.523644] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 678.524015] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.524207] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.524390] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 678.524718] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3b2a49f-669f-4e26-97ca-fd31a54c447c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.527668] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d14517c-543f-4386-b05f-8545571f043c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.559898] env[61991]: INFO nova.compute.manager [-] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Took 1.28 seconds to deallocate network for instance. [ 678.562846] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbdb29e-0aa3-436e-833d-1f7c4f5b0ec9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.565593] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 678.565768] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 678.571433] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b775e1e8-576f-456f-b179-9963704820ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.580310] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82aed2d-4823-4373-a0e4-2b15f957724f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.584829] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 678.584829] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5214b6b9-00c0-8a7f-5b46-1ae9183b08f8" [ 678.584829] env[61991]: _type = "Task" [ 678.584829] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.598299] env[61991]: DEBUG nova.compute.provider_tree [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.606555] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5214b6b9-00c0-8a7f-5b46-1ae9183b08f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008355} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.607909] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bbd7e54-b266-4fb0-980f-59c47eef8926 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.613916] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 678.613916] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52877ca9-a136-3aca-7dea-c4ffd73b8b4d" [ 678.613916] env[61991]: _type = "Task" [ 678.613916] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.623188] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52877ca9-a136-3aca-7dea-c4ffd73b8b4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.779751] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d59491d9-8d6e-4e10-975f-00eb18cf43c6 tempest-ServersTestFqdnHostnames-1906774487 tempest-ServersTestFqdnHostnames-1906774487-project-member] Lock "84b50fb2-7338-4da7-a1cb-d638c421a36d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.085s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.827430] env[61991]: DEBUG nova.network.neutron [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updated VIF entry in instance network info cache for port ea0fc9ad-ea54-4ccb-bce4-3c95855ec049. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 678.827562] env[61991]: DEBUG nova.network.neutron [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updating instance_info_cache with network_info: [{"id": "ea0fc9ad-ea54-4ccb-bce4-3c95855ec049", "address": "fa:16:3e:08:84:d5", "network": {"id": "0fccf457-0b11-4a14-a61a-493e55641197", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1825044270-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "403c79445dbe46609a260dbbf92386b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea0fc9ad-ea", "ovs_interfaceid": "ea0fc9ad-ea54-4ccb-bce4-3c95855ec049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.910466] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.913206] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "a2606541-ff39-4d03-bd19-d0ee8e035c08" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.913845] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.913845] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "a2606541-ff39-4d03-bd19-d0ee8e035c08-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.913987] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.914217] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.917713] env[61991]: INFO nova.compute.manager [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Terminating instance [ 678.919409] env[61991]: DEBUG nova.compute.manager [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.921272] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 678.921272] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4132ab0c-462b-436e-948b-40cf8b5b1184 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.928420] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 678.928885] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fa37e29-a68b-47fd-bc2c-db9e3584cff6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.935378] env[61991]: DEBUG oslo_vmware.api [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 678.935378] env[61991]: value = "task-1129374" [ 678.935378] env[61991]: _type = "Task" [ 678.935378] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.943734] env[61991]: DEBUG oslo_vmware.api [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129374, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.075725] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.103865] env[61991]: DEBUG nova.scheduler.client.report [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.127324] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52877ca9-a136-3aca-7dea-c4ffd73b8b4d, 'name': SearchDatastore_Task, 'duration_secs': 0.008723} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.131447] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.131723] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 9d88002b-d5f0-4669-9a34-bfbf8f9169d3/9d88002b-d5f0-4669-9a34-bfbf8f9169d3.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 679.131983] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4d59836-41c3-4740-a17f-2fd1b764f708 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.142149] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 679.142149] env[61991]: value = "task-1129375" [ 679.142149] env[61991]: _type = "Task" [ 679.142149] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.153499] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129375, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.282874] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.329789] env[61991]: DEBUG oslo_concurrency.lockutils [req-8af7a2be-667e-4cca-9e68-cebee239fda4 req-652a34da-a267-4897-bc31-a59ca4b2ee31 service nova] Releasing lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.441414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.449994] env[61991]: DEBUG oslo_vmware.api [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129374, 'name': PowerOffVM_Task, 'duration_secs': 0.165339} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.450342] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 679.450529] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 679.450909] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a95d5b26-f71f-447a-b915-2a59a2050931 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.543403] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 679.543722] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 679.543979] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Deleting the datastore file [datastore1] a2606541-ff39-4d03-bd19-d0ee8e035c08 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 679.544312] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa44f232-c52e-4c7e-a87c-eb86e3d09f1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.553307] env[61991]: DEBUG oslo_vmware.api [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for the task: (returnval){ [ 679.553307] env[61991]: value = "task-1129377" [ 679.553307] env[61991]: _type = "Task" [ 679.553307] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.563331] env[61991]: DEBUG oslo_vmware.api [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129377, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.611453] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.612219] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.615878] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.282s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.617558] env[61991]: INFO nova.compute.claims [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.654351] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129375, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463935} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.654672] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 9d88002b-d5f0-4669-9a34-bfbf8f9169d3/9d88002b-d5f0-4669-9a34-bfbf8f9169d3.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 679.654901] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 679.655173] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e588c1e8-98f6-4771-9f97-2414ee1ac1b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.661821] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 679.661821] env[61991]: value = "task-1129378" [ 679.661821] env[61991]: _type = "Task" [ 679.661821] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.671215] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129378, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.814270] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.063858] env[61991]: DEBUG oslo_vmware.api [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Task: {'id': task-1129377, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181892} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.063858] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 680.064049] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 680.064229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 680.064398] env[61991]: INFO nova.compute.manager [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Took 1.14 seconds to destroy the instance on the hypervisor. [ 680.064828] env[61991]: DEBUG oslo.service.loopingcall [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.064828] env[61991]: DEBUG nova.compute.manager [-] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.064914] env[61991]: DEBUG nova.network.neutron [-] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.117276] env[61991]: DEBUG nova.compute.utils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.118781] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.118954] env[61991]: DEBUG nova.network.neutron [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 680.171571] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129378, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086609} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.171950] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 680.172970] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466533d2-5bcc-4743-873f-f1d92c43b108 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.195197] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 9d88002b-d5f0-4669-9a34-bfbf8f9169d3/9d88002b-d5f0-4669-9a34-bfbf8f9169d3.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 680.195504] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f014a255-960c-4503-9fda-21e10d3eac2d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.214382] env[61991]: DEBUG nova.policy [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc2e1f75d6bd44df96ad5b229d883b31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25647cd0715f490092623f4ead833738', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.221763] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 680.221763] env[61991]: value = "task-1129379" [ 680.221763] env[61991]: _type = "Task" [ 680.221763] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.229710] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129379, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.537778] env[61991]: DEBUG nova.compute.manager [req-2846b7c3-7065-45ad-a984-c0ac42a3345d req-473da970-286d-45a7-9173-72c6e437d36a service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Received event network-vif-deleted-4b335179-f726-4551-a93c-01b6472885e1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.537980] env[61991]: INFO nova.compute.manager [req-2846b7c3-7065-45ad-a984-c0ac42a3345d req-473da970-286d-45a7-9173-72c6e437d36a service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Neutron deleted interface 4b335179-f726-4551-a93c-01b6472885e1; detaching it from the instance and deleting it from the info cache [ 680.538091] env[61991]: DEBUG nova.network.neutron [req-2846b7c3-7065-45ad-a984-c0ac42a3345d req-473da970-286d-45a7-9173-72c6e437d36a service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.622190] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.734476] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129379, 'name': ReconfigVM_Task, 'duration_secs': 0.494066} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.734840] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 9d88002b-d5f0-4669-9a34-bfbf8f9169d3/9d88002b-d5f0-4669-9a34-bfbf8f9169d3.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.735777] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3bc07bf9-3622-4f7e-892f-94e08549d2ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.742464] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 680.742464] env[61991]: value = "task-1129380" [ 680.742464] env[61991]: _type = "Task" [ 680.742464] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.756529] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129380, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.785660] env[61991]: DEBUG nova.network.neutron [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Successfully created port: 6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.007497] env[61991]: DEBUG nova.network.neutron [-] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.046180] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3690e3dd-77c8-4778-8f92-02c06e2c88fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.052133] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bafbd173-3a57-44ac-a936-0ed04fe5c2f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.060026] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4c69be-7ee2-4447-af89-47fa808668fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.066509] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0218dff7-f28c-4334-8676-087bec57571b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.110927] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398ef4ee-ed49-4e37-ad06-410283951630 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.116685] env[61991]: DEBUG nova.compute.manager [req-2846b7c3-7065-45ad-a984-c0ac42a3345d req-473da970-286d-45a7-9173-72c6e437d36a service nova] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Detach interface failed, port_id=4b335179-f726-4551-a93c-01b6472885e1, reason: Instance a2606541-ff39-4d03-bd19-d0ee8e035c08 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 681.119633] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7fca50-640b-400e-889f-9674cb128a8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.138205] env[61991]: DEBUG nova.compute.provider_tree [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.258040] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129380, 'name': Rename_Task, 'duration_secs': 0.144402} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.258040] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 681.258040] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59c14bd1-0b05-4f09-96e3-8ff36e827895 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.265939] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 681.265939] env[61991]: value = "task-1129381" [ 681.265939] env[61991]: _type = "Task" [ 681.265939] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.278277] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129381, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.512249] env[61991]: INFO nova.compute.manager [-] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Took 1.45 seconds to deallocate network for instance. [ 681.644531] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.645248] env[61991]: DEBUG nova.scheduler.client.report [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.687244] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.687620] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.687712] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.687819] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.687963] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.688205] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.688419] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.688575] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.688739] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.688896] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.689087] env[61991]: DEBUG nova.virt.hardware [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.689947] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65192b71-de1a-4af9-a5ab-9ab4974f419b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.699349] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5224137d-567b-4953-9b02-529717aed5af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.777841] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129381, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.021438] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.151320] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.152017] env[61991]: DEBUG nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.155393] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.135s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.158489] env[61991]: INFO nova.compute.claims [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.278273] env[61991]: DEBUG oslo_vmware.api [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129381, 'name': PowerOnVM_Task, 'duration_secs': 1.011442} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.278816] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 682.279158] env[61991]: INFO nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Took 7.15 seconds to spawn the instance on the hypervisor. [ 682.279853] env[61991]: DEBUG nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 682.281182] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64985c8b-2594-439e-9988-1b724f098e90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.669958] env[61991]: DEBUG nova.compute.utils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.676521] env[61991]: DEBUG nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 682.761992] env[61991]: DEBUG nova.compute.manager [req-c2f998e3-bf78-443d-b86c-c5a74db7d380 req-0dd61d4f-0a90-4058-b0ae-41f5b9b623e0 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Received event network-vif-plugged-6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.762241] env[61991]: DEBUG oslo_concurrency.lockutils [req-c2f998e3-bf78-443d-b86c-c5a74db7d380 req-0dd61d4f-0a90-4058-b0ae-41f5b9b623e0 service nova] Acquiring lock "812d3491-0598-4171-b905-5304ebf6ae37-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.762454] env[61991]: DEBUG oslo_concurrency.lockutils [req-c2f998e3-bf78-443d-b86c-c5a74db7d380 req-0dd61d4f-0a90-4058-b0ae-41f5b9b623e0 service nova] Lock "812d3491-0598-4171-b905-5304ebf6ae37-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.762935] env[61991]: DEBUG oslo_concurrency.lockutils [req-c2f998e3-bf78-443d-b86c-c5a74db7d380 req-0dd61d4f-0a90-4058-b0ae-41f5b9b623e0 service nova] Lock "812d3491-0598-4171-b905-5304ebf6ae37-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.762935] env[61991]: DEBUG nova.compute.manager [req-c2f998e3-bf78-443d-b86c-c5a74db7d380 req-0dd61d4f-0a90-4058-b0ae-41f5b9b623e0 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] No waiting events found dispatching network-vif-plugged-6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 682.762935] env[61991]: WARNING nova.compute.manager [req-c2f998e3-bf78-443d-b86c-c5a74db7d380 req-0dd61d4f-0a90-4058-b0ae-41f5b9b623e0 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Received unexpected event network-vif-plugged-6d378d4e-f7a2-47e8-a0f1-135318bd5381 for instance with vm_state building and task_state spawning. [ 682.806290] env[61991]: INFO nova.compute.manager [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Took 43.26 seconds to build instance. [ 683.073460] env[61991]: DEBUG nova.network.neutron [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Successfully updated port: 6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 683.176763] env[61991]: DEBUG nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.311086] env[61991]: DEBUG oslo_concurrency.lockutils [None req-854b6b0b-aa85-424c-a300-d102d82ef61b tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.262s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.549107] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a7a8a4-a29f-43e6-9dab-8736c634a56e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.557157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a452cff3-95ac-42c2-a322-c3a4e83bf8c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.588597] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "refresh_cache-812d3491-0598-4171-b905-5304ebf6ae37" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.588597] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "refresh_cache-812d3491-0598-4171-b905-5304ebf6ae37" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.588739] env[61991]: DEBUG nova.network.neutron [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 683.594125] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1f3410-82b0-4602-8915-baf3f24f08d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.602801] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e81199-0747-4d65-acef-cbcefecc0c57 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.622118] env[61991]: DEBUG nova.compute.provider_tree [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.819707] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.126409] env[61991]: DEBUG nova.scheduler.client.report [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.151897] env[61991]: DEBUG nova.network.neutron [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.191227] env[61991]: DEBUG nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.234021] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.234021] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.234021] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.234021] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.234306] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.234306] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.234306] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.234306] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.234306] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.234507] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.234507] env[61991]: DEBUG nova.virt.hardware [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.234982] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ad1bbc-9d26-4779-b8d2-255aa828cccf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.246600] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddaa0caa-0e10-4dd8-aae1-1848fa43169b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.264118] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 684.269269] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Creating folder: Project (93c3c8a544e248a5aa514f95d4dc925c). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 684.269506] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a9eecb4-7551-4dbd-aae9-4cb02a4bf3b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.282755] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Created folder: Project (93c3c8a544e248a5aa514f95d4dc925c) in parent group-v246753. [ 684.282964] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Creating folder: Instances. Parent ref: group-v246783. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 684.283207] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44e2343f-e4a9-485e-a9f1-a098ffd0a092 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.292444] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Created folder: Instances in parent group-v246783. [ 684.292698] env[61991]: DEBUG oslo.service.loopingcall [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.292891] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 684.293427] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f17b35fa-4bed-45fe-be33-70dab7a9619a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.319037] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 684.319037] env[61991]: value = "task-1129384" [ 684.319037] env[61991]: _type = "Task" [ 684.319037] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.329264] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129384, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.352896] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.361203] env[61991]: DEBUG nova.network.neutron [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Updating instance_info_cache with network_info: [{"id": "6d378d4e-f7a2-47e8-a0f1-135318bd5381", "address": "fa:16:3e:bd:b6:f4", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d378d4e-f7", "ovs_interfaceid": "6d378d4e-f7a2-47e8-a0f1-135318bd5381", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.634159] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.634796] env[61991]: DEBUG nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 684.643428] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.637s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.643428] env[61991]: DEBUG nova.objects.instance [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lazy-loading 'resources' on Instance uuid 2866aec9-56fd-4e77-b651-9d85783fb94f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.685886] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.686394] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.714784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.715059] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.746315] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "671e0197-4f96-4660-be99-ea1d1c0588e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.746534] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.809054] env[61991]: DEBUG nova.compute.manager [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Received event network-changed-6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.809164] env[61991]: DEBUG nova.compute.manager [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Refreshing instance network info cache due to event network-changed-6d378d4e-f7a2-47e8-a0f1-135318bd5381. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 684.811020] env[61991]: DEBUG oslo_concurrency.lockutils [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] Acquiring lock "refresh_cache-812d3491-0598-4171-b905-5304ebf6ae37" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.829730] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129384, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.864172] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "refresh_cache-812d3491-0598-4171-b905-5304ebf6ae37" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.864465] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Instance network_info: |[{"id": "6d378d4e-f7a2-47e8-a0f1-135318bd5381", "address": "fa:16:3e:bd:b6:f4", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d378d4e-f7", "ovs_interfaceid": "6d378d4e-f7a2-47e8-a0f1-135318bd5381", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 684.864806] env[61991]: DEBUG oslo_concurrency.lockutils [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] Acquired lock "refresh_cache-812d3491-0598-4171-b905-5304ebf6ae37" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.864944] env[61991]: DEBUG nova.network.neutron [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Refreshing network info cache for port 6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 684.866216] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:b6:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc845e3-654b-43c6-acea-dde1084f0ad0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d378d4e-f7a2-47e8-a0f1-135318bd5381', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 684.873567] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Creating folder: Project (25647cd0715f490092623f4ead833738). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 684.874353] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02dbce59-d96e-4590-bbf8-266983f4d866 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.885433] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Created folder: Project (25647cd0715f490092623f4ead833738) in parent group-v246753. [ 684.885639] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Creating folder: Instances. Parent ref: group-v246786. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 684.886254] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-431db334-304b-4cc1-9bdc-86a9b2a51542 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.896191] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Created folder: Instances in parent group-v246786. [ 684.896343] env[61991]: DEBUG oslo.service.loopingcall [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.896525] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 684.896721] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ff9e877-647a-42fe-8134-227b7802e94c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.919169] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 684.919169] env[61991]: value = "task-1129387" [ 684.919169] env[61991]: _type = "Task" [ 684.919169] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.928013] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129387, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.147322] env[61991]: DEBUG nova.compute.utils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.150352] env[61991]: DEBUG nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 685.335586] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129384, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.434950] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129387, 'name': CreateVM_Task, 'duration_secs': 0.404569} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.434950] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 685.436246] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.436425] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.436847] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 685.439440] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba1e0fde-acd5-45fc-acfd-3d2e6ef68b5b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.445950] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 685.445950] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c93026-3c59-7640-4278-36ae554916e1" [ 685.445950] env[61991]: _type = "Task" [ 685.445950] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.455639] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c93026-3c59-7640-4278-36ae554916e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.598669] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29bd799-81d7-46a4-8aea-f649870dd51e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.607027] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae82bcd-92cd-46ce-9ad1-a0c35fce396f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.645154] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdc0f86-bcc6-43c6-82bf-79a0015ce523 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.652955] env[61991]: DEBUG nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.657024] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10d61c9-13bc-4b1b-84bf-b6a588e06754 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.673309] env[61991]: DEBUG nova.compute.provider_tree [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.688968] env[61991]: DEBUG nova.network.neutron [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Updated VIF entry in instance network info cache for port 6d378d4e-f7a2-47e8-a0f1-135318bd5381. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 685.689472] env[61991]: DEBUG nova.network.neutron [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Updating instance_info_cache with network_info: [{"id": "6d378d4e-f7a2-47e8-a0f1-135318bd5381", "address": "fa:16:3e:bd:b6:f4", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d378d4e-f7", "ovs_interfaceid": "6d378d4e-f7a2-47e8-a0f1-135318bd5381", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.764221] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "d6a2d226-5097-4461-b2ce-fa698f1066d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.764452] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.831400] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129384, 'name': CreateVM_Task, 'duration_secs': 1.297378} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.831692] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 685.832112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.959441] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c93026-3c59-7640-4278-36ae554916e1, 'name': SearchDatastore_Task, 'duration_secs': 0.00989} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.959441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.959441] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 685.959441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.959624] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.959624] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 685.959624] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.959624] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 685.959624] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee5c5b1b-6256-4310-bee4-fe75c02187e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.961374] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1f775c8-0d44-4de9-9e1b-126dda2767df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.967096] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 685.967096] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d92f0b-7cd6-9a72-dcef-bc0f3b98d608" [ 685.967096] env[61991]: _type = "Task" [ 685.967096] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.972584] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 685.972790] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 685.973922] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6afb559e-1c30-4191-9167-94f8bf6be52c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.981569] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d92f0b-7cd6-9a72-dcef-bc0f3b98d608, 'name': SearchDatastore_Task, 'duration_secs': 0.009725} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.982252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.982310] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 685.982512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.984276] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 685.984276] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b35c09-cec8-0cff-7cd3-0c6b411ae1a2" [ 685.984276] env[61991]: _type = "Task" [ 685.984276] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.991840] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b35c09-cec8-0cff-7cd3-0c6b411ae1a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.034122] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.034405] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.034618] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.034802] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.034965] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.037100] env[61991]: INFO nova.compute.manager [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Terminating instance [ 686.038833] env[61991]: DEBUG nova.compute.manager [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.039034] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 686.039857] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee7d2ea-edd4-44ec-baa6-bc339324d7b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.047237] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 686.048088] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ab3b67e-a67e-4c14-abae-cacc85710959 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.053441] env[61991]: DEBUG oslo_vmware.api [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 686.053441] env[61991]: value = "task-1129388" [ 686.053441] env[61991]: _type = "Task" [ 686.053441] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.060997] env[61991]: DEBUG oslo_vmware.api [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.175533] env[61991]: DEBUG nova.scheduler.client.report [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.192499] env[61991]: DEBUG oslo_concurrency.lockutils [req-33e9edc7-1076-42ad-90e0-5aeaa3eaa4d9 req-bf8c1932-d60a-4fd1-a711-fc47df64850c service nova] Releasing lock "refresh_cache-812d3491-0598-4171-b905-5304ebf6ae37" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.494866] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b35c09-cec8-0cff-7cd3-0c6b411ae1a2, 'name': SearchDatastore_Task, 'duration_secs': 0.008761} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.496200] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d582eb8-6c26-4dd2-bc50-fd3bfc846707 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.500659] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 686.500659] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525558ad-10f2-9b9c-7a8a-3af5a8ffe2d9" [ 686.500659] env[61991]: _type = "Task" [ 686.500659] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.509599] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525558ad-10f2-9b9c-7a8a-3af5a8ffe2d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.563268] env[61991]: DEBUG oslo_vmware.api [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129388, 'name': PowerOffVM_Task, 'duration_secs': 0.202915} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.563543] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 686.563731] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 686.563985] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed95a264-2805-4c78-9a15-24c888fe5be0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.625067] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 686.625455] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 686.625775] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Deleting the datastore file [datastore1] 9d88002b-d5f0-4669-9a34-bfbf8f9169d3 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 686.626152] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9a6281d-3d1e-4dc7-910f-a3ab50ad4278 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.633514] env[61991]: DEBUG oslo_vmware.api [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for the task: (returnval){ [ 686.633514] env[61991]: value = "task-1129390" [ 686.633514] env[61991]: _type = "Task" [ 686.633514] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.641727] env[61991]: DEBUG oslo_vmware.api [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.668319] env[61991]: DEBUG nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.680605] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.682722] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.960s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.684302] env[61991]: INFO nova.compute.claims [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.703045] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.703289] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.703492] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.703610] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.703797] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.704146] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.704146] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.704283] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.704446] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.704603] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.704766] env[61991]: DEBUG nova.virt.hardware [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.705617] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d6a65d-0b3a-4546-839e-1a310f4e15fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.709420] env[61991]: INFO nova.scheduler.client.report [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Deleted allocations for instance 2866aec9-56fd-4e77-b651-9d85783fb94f [ 686.716376] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542b4144-3190-4e1e-8667-5aed280e8569 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.730559] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.736291] env[61991]: DEBUG oslo.service.loopingcall [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.737185] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 686.737391] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d132b44-6156-40b1-a159-d3b7310d88f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.754552] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.754552] env[61991]: value = "task-1129391" [ 686.754552] env[61991]: _type = "Task" [ 686.754552] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.763538] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129391, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.011531] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525558ad-10f2-9b9c-7a8a-3af5a8ffe2d9, 'name': SearchDatastore_Task, 'duration_secs': 0.009096} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.011826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.012014] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 812d3491-0598-4171-b905-5304ebf6ae37/812d3491-0598-4171-b905-5304ebf6ae37.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 687.012292] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.013023] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 687.013023] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2b40389-d3e3-4793-928a-3c4f50ea77d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.014567] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1708a779-5b0b-44b7-b3b9-b28243566a67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.021643] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 687.021643] env[61991]: value = "task-1129392" [ 687.021643] env[61991]: _type = "Task" [ 687.021643] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.022644] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 687.022882] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 687.026237] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9706a79f-c4ff-470c-a7ea-f3f22ffb4f25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.033026] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.034201] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 687.034201] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f08e07-3974-c0f4-e6cb-fcb733656a98" [ 687.034201] env[61991]: _type = "Task" [ 687.034201] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.042335] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f08e07-3974-c0f4-e6cb-fcb733656a98, 'name': SearchDatastore_Task, 'duration_secs': 0.00867} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.043088] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54123737-79f6-421b-898a-611ee902e11a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.047618] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 687.047618] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d63263-fa95-463e-4ad3-b8985cb54cf9" [ 687.047618] env[61991]: _type = "Task" [ 687.047618] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.054996] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d63263-fa95-463e-4ad3-b8985cb54cf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.143301] env[61991]: DEBUG oslo_vmware.api [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Task: {'id': task-1129390, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136264} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.143566] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 687.143816] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 687.144057] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.144212] env[61991]: INFO nova.compute.manager [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 687.144469] env[61991]: DEBUG oslo.service.loopingcall [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.144731] env[61991]: DEBUG nova.compute.manager [-] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.144843] env[61991]: DEBUG nova.network.neutron [-] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.224153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b5849fc1-7ca3-4efb-8dd8-fbaf71f3c2b7 tempest-ServerAddressesNegativeTestJSON-1851156385 tempest-ServerAddressesNegativeTestJSON-1851156385-project-member] Lock "2866aec9-56fd-4e77-b651-9d85783fb94f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.124s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.266770] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129391, 'name': CreateVM_Task, 'duration_secs': 0.256301} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.267018] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 687.267385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.267540] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.267852] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 687.268121] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94da5cfd-543b-4e82-b71a-efc0d28a672f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.272818] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 687.272818] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528bf41f-ca0c-385b-fe64-fa56133d1ea7" [ 687.272818] env[61991]: _type = "Task" [ 687.272818] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.281098] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528bf41f-ca0c-385b-fe64-fa56133d1ea7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.485243] env[61991]: DEBUG nova.compute.manager [req-bde2f7b0-73a7-4701-af58-0f9345ffdcee req-63de072c-71d2-47a8-885f-94df8f501379 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Received event network-vif-deleted-ea0fc9ad-ea54-4ccb-bce4-3c95855ec049 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.485463] env[61991]: INFO nova.compute.manager [req-bde2f7b0-73a7-4701-af58-0f9345ffdcee req-63de072c-71d2-47a8-885f-94df8f501379 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Neutron deleted interface ea0fc9ad-ea54-4ccb-bce4-3c95855ec049; detaching it from the instance and deleting it from the info cache [ 687.485636] env[61991]: DEBUG nova.network.neutron [req-bde2f7b0-73a7-4701-af58-0f9345ffdcee req-63de072c-71d2-47a8-885f-94df8f501379 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.532398] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457106} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.532398] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 812d3491-0598-4171-b905-5304ebf6ae37/812d3491-0598-4171-b905-5304ebf6ae37.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 687.532398] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 687.532398] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8924040f-e209-43da-8a1f-9a3656fd4ee5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.539723] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 687.539723] env[61991]: value = "task-1129393" [ 687.539723] env[61991]: _type = "Task" [ 687.539723] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.548659] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.556923] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d63263-fa95-463e-4ad3-b8985cb54cf9, 'name': SearchDatastore_Task, 'duration_secs': 0.0079} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.557224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.557496] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] acef131a-4263-42ad-964e-bb1e0cb21eec/acef131a-4263-42ad-964e-bb1e0cb21eec.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 687.557774] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74d05220-c491-4d5b-b468-6739b3dc3c28 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.564437] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 687.564437] env[61991]: value = "task-1129394" [ 687.564437] env[61991]: _type = "Task" [ 687.564437] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.572519] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129394, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.789959] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528bf41f-ca0c-385b-fe64-fa56133d1ea7, 'name': SearchDatastore_Task, 'duration_secs': 0.05755} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.790576] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.790927] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 687.791290] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.791552] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.791841] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 687.792349] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9aabfdd-42e3-4f5d-ac18-f0ae35402b4d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.809197] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 687.809197] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 687.809197] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad7bab4-d9a7-4bb8-88c0-a46571736f23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.823189] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 687.823189] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52921611-0125-981f-45f4-29875fa747ff" [ 687.823189] env[61991]: _type = "Task" [ 687.823189] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.833418] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52921611-0125-981f-45f4-29875fa747ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.958643] env[61991]: DEBUG nova.network.neutron [-] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.987940] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93b689d4-4a93-4d1b-9b7c-7d2d9dde1618 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.996511] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d104166-482c-43be-a08b-7e469aef0351 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.024400] env[61991]: DEBUG nova.compute.manager [req-bde2f7b0-73a7-4701-af58-0f9345ffdcee req-63de072c-71d2-47a8-885f-94df8f501379 service nova] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Detach interface failed, port_id=ea0fc9ad-ea54-4ccb-bce4-3c95855ec049, reason: Instance 9d88002b-d5f0-4669-9a34-bfbf8f9169d3 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 688.049310] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063023} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.051629] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 688.052642] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b19b78-0c7a-485b-b8f6-198cfa2cd5d4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.073855] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 812d3491-0598-4171-b905-5304ebf6ae37/812d3491-0598-4171-b905-5304ebf6ae37.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 688.078961] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b270f96a-f94e-44e1-b8e7-bcff63bf0b6c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.098460] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129394, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427844} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.099815] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] acef131a-4263-42ad-964e-bb1e0cb21eec/acef131a-4263-42ad-964e-bb1e0cb21eec.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 688.100785] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 688.100785] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 688.100785] env[61991]: value = "task-1129395" [ 688.100785] env[61991]: _type = "Task" [ 688.100785] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.100785] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2f99819-2acd-4dd4-b5a7-9b472ceeea1a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.112396] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129395, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.113636] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 688.113636] env[61991]: value = "task-1129396" [ 688.113636] env[61991]: _type = "Task" [ 688.113636] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.124837] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.159189] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78eb691-e2f7-4a32-8526-531e28ae3ec4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.167834] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fc5791-e4c9-4eac-9df4-400889dea2a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.197898] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852a7328-638e-4e60-85df-bd0b6a16e22d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.205380] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d01bd74-3e8e-4e37-ab04-5a4b23cd567b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.219502] env[61991]: DEBUG nova.compute.provider_tree [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.333807] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52921611-0125-981f-45f4-29875fa747ff, 'name': SearchDatastore_Task, 'duration_secs': 0.047488} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.334430] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-086c580d-51f2-4944-9879-5efc4f363411 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.340829] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 688.340829] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52946b28-90ef-f410-9d13-0426e341166e" [ 688.340829] env[61991]: _type = "Task" [ 688.340829] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.350904] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52946b28-90ef-f410-9d13-0426e341166e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.464401] env[61991]: INFO nova.compute.manager [-] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Took 1.32 seconds to deallocate network for instance. [ 688.612939] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129395, 'name': ReconfigVM_Task, 'duration_secs': 0.266733} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.613413] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 812d3491-0598-4171-b905-5304ebf6ae37/812d3491-0598-4171-b905-5304ebf6ae37.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 688.614403] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38333e49-a709-453b-935a-32c803efc2f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.623721] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062486} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.624916] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 688.625266] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 688.625266] env[61991]: value = "task-1129397" [ 688.625266] env[61991]: _type = "Task" [ 688.625266] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.625929] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9c98e7-1ce6-4a1d-96aa-c306a361c048 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.635979] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129397, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.651110] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] acef131a-4263-42ad-964e-bb1e0cb21eec/acef131a-4263-42ad-964e-bb1e0cb21eec.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 688.651406] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38541fbe-6640-441c-92b2-c0dfdf6fb1a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.670768] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 688.670768] env[61991]: value = "task-1129398" [ 688.670768] env[61991]: _type = "Task" [ 688.670768] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.678717] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129398, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.723162] env[61991]: DEBUG nova.scheduler.client.report [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.851265] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52946b28-90ef-f410-9d13-0426e341166e, 'name': SearchDatastore_Task, 'duration_secs': 0.01064} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.851577] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.851868] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 688.852142] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ef9eea0-4ae7-4955-8f07-6951695506fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.859132] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 688.859132] env[61991]: value = "task-1129399" [ 688.859132] env[61991]: _type = "Task" [ 688.859132] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.867390] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129399, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.971402] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.139750] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129397, 'name': Rename_Task, 'duration_secs': 0.178224} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.140118] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 689.140265] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94654d2c-8d8b-40f8-860e-df8d44ee7336 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.148639] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 689.148639] env[61991]: value = "task-1129400" [ 689.148639] env[61991]: _type = "Task" [ 689.148639] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.158435] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.182311] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129398, 'name': ReconfigVM_Task, 'duration_secs': 0.265114} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.182601] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Reconfigured VM instance instance-00000028 to attach disk [datastore2] acef131a-4263-42ad-964e-bb1e0cb21eec/acef131a-4263-42ad-964e-bb1e0cb21eec.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 689.183822] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-579155ea-76b0-4249-baae-185c8403f4a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.189306] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 689.189306] env[61991]: value = "task-1129401" [ 689.189306] env[61991]: _type = "Task" [ 689.189306] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.200546] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129401, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.228729] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.229346] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.232557] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.393s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.234327] env[61991]: INFO nova.compute.claims [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.369744] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129399, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496021} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.370063] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 689.370280] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 689.370588] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6aa9a0ab-5135-4cd1-92f2-cca2f0d46ca8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.377631] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 689.377631] env[61991]: value = "task-1129402" [ 689.377631] env[61991]: _type = "Task" [ 689.377631] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.385498] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129402, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.660078] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129400, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.700564] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129401, 'name': Rename_Task, 'duration_secs': 0.21279} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.701670] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 689.701670] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3265c240-425d-45f7-bcdc-f04ceab4303a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.708520] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 689.708520] env[61991]: value = "task-1129403" [ 689.708520] env[61991]: _type = "Task" [ 689.708520] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.718926] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.739892] env[61991]: DEBUG nova.compute.utils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.743864] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.744037] env[61991]: DEBUG nova.network.neutron [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 689.810402] env[61991]: DEBUG nova.policy [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8c995dbd884e95bd04e034e2999a56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e4cb3d1518d498a8cdc2aee3acb90cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 689.886962] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129402, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073584} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.887260] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 689.888051] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72db0f67-6874-4eae-a370-87f5ec3a16c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.907804] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 689.907804] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f515d8e4-58d9-4f8a-b6d6-0c9ac5a8f413 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.927350] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 689.927350] env[61991]: value = "task-1129404" [ 689.927350] env[61991]: _type = "Task" [ 689.927350] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.935508] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129404, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.159598] env[61991]: DEBUG oslo_vmware.api [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129400, 'name': PowerOnVM_Task, 'duration_secs': 0.606084} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.159934] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 690.160115] env[61991]: INFO nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Took 8.52 seconds to spawn the instance on the hypervisor. [ 690.160312] env[61991]: DEBUG nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 690.161136] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b5b080-3a3c-43b5-81a7-737e48b89255 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.164750] env[61991]: DEBUG nova.network.neutron [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Successfully created port: 87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.219538] env[61991]: DEBUG oslo_vmware.api [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129403, 'name': PowerOnVM_Task, 'duration_secs': 0.481969} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.219785] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 690.219975] env[61991]: INFO nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Took 6.03 seconds to spawn the instance on the hypervisor. [ 690.220243] env[61991]: DEBUG nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 690.221012] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca1492a-1d27-4289-a6c4-6433df69b008 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.244230] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.439078] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129404, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.603483] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b690b55f-48d2-4b94-950a-294b40aa50e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.611192] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba5de1a-53c5-477a-b1fc-e40041e910a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.642282] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37fe25c9-322a-42b4-8e01-4de439557270 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.650345] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9b81d4-2b7e-40ea-8721-2ffb51d03cd9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.663642] env[61991]: DEBUG nova.compute.provider_tree [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.680354] env[61991]: INFO nova.compute.manager [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Took 29.35 seconds to build instance. [ 690.735945] env[61991]: INFO nova.compute.manager [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Took 27.42 seconds to build instance. [ 690.939350] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129404, 'name': ReconfigVM_Task, 'duration_secs': 0.777184} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.939616] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Reconfigured VM instance instance-00000029 to attach disk [datastore2] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 690.940222] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2282ccca-3ad7-4f4a-af3c-22d6042b8c33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.946392] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 690.946392] env[61991]: value = "task-1129405" [ 690.946392] env[61991]: _type = "Task" [ 690.946392] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.955344] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129405, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.170508] env[61991]: DEBUG nova.scheduler.client.report [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.182461] env[61991]: DEBUG oslo_concurrency.lockutils [None req-920257e7-7735-4857-80bb-c42cff1a30f4 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "812d3491-0598-4171-b905-5304ebf6ae37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.531s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.238297] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a2687d8e-ce3a-42b2-a29c-b7ea78a81002 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "acef131a-4263-42ad-964e-bb1e0cb21eec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.061s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.253384] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 691.276916] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.277182] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.277335] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.277512] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.277654] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.277796] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.277998] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.278176] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.278340] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.278497] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.278662] env[61991]: DEBUG nova.virt.hardware [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.279539] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac19261-19b1-4dee-9de1-4daaf50ef75b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.287836] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176adab9-d13e-4d01-8ab3-558325283279 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.456655] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129405, 'name': Rename_Task, 'duration_secs': 0.173005} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.457049] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 691.457323] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fe155db-098c-4380-8435-8b5118ea2c6c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.463559] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 691.463559] env[61991]: value = "task-1129406" [ 691.463559] env[61991]: _type = "Task" [ 691.463559] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.471748] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.675647] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.676220] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.683142] env[61991]: DEBUG nova.compute.manager [req-306712c4-ecfa-4026-9815-761f56746cec req-59ea135a-344c-4051-9cf3-d39847e8c2d9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Received event network-vif-plugged-87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.683142] env[61991]: DEBUG oslo_concurrency.lockutils [req-306712c4-ecfa-4026-9815-761f56746cec req-59ea135a-344c-4051-9cf3-d39847e8c2d9 service nova] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.683142] env[61991]: DEBUG oslo_concurrency.lockutils [req-306712c4-ecfa-4026-9815-761f56746cec req-59ea135a-344c-4051-9cf3-d39847e8c2d9 service nova] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.683142] env[61991]: DEBUG oslo_concurrency.lockutils [req-306712c4-ecfa-4026-9815-761f56746cec req-59ea135a-344c-4051-9cf3-d39847e8c2d9 service nova] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.683142] env[61991]: DEBUG nova.compute.manager [req-306712c4-ecfa-4026-9815-761f56746cec req-59ea135a-344c-4051-9cf3-d39847e8c2d9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] No waiting events found dispatching network-vif-plugged-87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 691.683482] env[61991]: WARNING nova.compute.manager [req-306712c4-ecfa-4026-9815-761f56746cec req-59ea135a-344c-4051-9cf3-d39847e8c2d9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Received unexpected event network-vif-plugged-87a47717-3b2c-4160-97b7-ef566ac540c0 for instance with vm_state building and task_state spawning. [ 691.683482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.254s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.683482] env[61991]: INFO nova.compute.claims [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.686012] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.740206] env[61991]: DEBUG nova.network.neutron [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Successfully updated port: 87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 691.740998] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.977113] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129406, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.182752] env[61991]: DEBUG nova.compute.utils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.184155] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 692.184325] env[61991]: DEBUG nova.network.neutron [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 692.210836] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.244715] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.244907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.245092] env[61991]: DEBUG nova.network.neutron [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.252804] env[61991]: DEBUG nova.policy [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2666f4b636d4e4d8682a5876772d381', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '518b363cb2db43c288d25c39a239c9c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.264382] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.475581] env[61991]: DEBUG oslo_vmware.api [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129406, 'name': PowerOnVM_Task, 'duration_secs': 0.745801} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.475907] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 692.476141] env[61991]: INFO nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Took 5.81 seconds to spawn the instance on the hypervisor. [ 692.476357] env[61991]: DEBUG nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 692.477232] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cac722-3bfc-42ca-9144-0e7e0b701780 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.558751] env[61991]: DEBUG nova.network.neutron [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Successfully created port: 863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.688221] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.785690] env[61991]: DEBUG nova.network.neutron [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.959081] env[61991]: DEBUG nova.network.neutron [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating instance_info_cache with network_info: [{"id": "87a47717-3b2c-4160-97b7-ef566ac540c0", "address": "fa:16:3e:e6:5f:4d", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87a47717-3b", "ovs_interfaceid": "87a47717-3b2c-4160-97b7-ef566ac540c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.984465] env[61991]: DEBUG nova.compute.manager [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 692.985451] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caf9b35-be80-4c9b-acd5-e8868abcc591 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.369441] env[61991]: INFO nova.compute.manager [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Took 29.36 seconds to build instance. [ 693.437396] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efe8403-3e93-4a1d-8a03-ec2fcdd5bd53 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.447490] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedccb01-5f61-4e70-9798-6fd9c55bbc20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.491886] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97e51fa-16c5-4d1b-bd4f-fafe04046f2c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.502569] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4b80d3-9f82-4361-86bb-9ced29e770f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.525218] env[61991]: DEBUG nova.compute.provider_tree [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.866386] env[61991]: INFO nova.compute.manager [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] instance snapshotting [ 693.868382] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.868482] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Instance network_info: |[{"id": "87a47717-3b2c-4160-97b7-ef566ac540c0", "address": "fa:16:3e:e6:5f:4d", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87a47717-3b", "ovs_interfaceid": "87a47717-3b2c-4160-97b7-ef566ac540c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 693.869181] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:5f:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87a47717-3b2c-4160-97b7-ef566ac540c0', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 693.876413] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Creating folder: Project (9e4cb3d1518d498a8cdc2aee3acb90cb). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.877149] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efd265e-e1d7-47dc-a40e-e633773e90f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.879727] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-293c78e1-b4de-4b55-966a-b956e2154b01 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.881516] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d26acef-59bf-4cf2-a109-d90b5e2f44dc tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "fe78a8de-6174-47b3-9271-171c4ba76068" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.166s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.885467] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.904825] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90410624-4b4b-41bc-9e94-540ef92cbe43 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.907609] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Created folder: Project (9e4cb3d1518d498a8cdc2aee3acb90cb) in parent group-v246753. [ 693.907801] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Creating folder: Instances. Parent ref: group-v246790. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.908248] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1486dca5-fc5b-4ef2-88a1-050d93a7b25a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.917581] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:18:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b41f5b31-f99f-43b3-b2c2-a0cd889c0027',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-424611864',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.917581] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.917581] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.917763] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.917763] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.917763] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.917763] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.917763] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.917960] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.917960] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.917960] env[61991]: DEBUG nova.virt.hardware [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.918655] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e234c691-9e0a-4f2d-959e-0c3cc4b46958 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.930239] env[61991]: DEBUG nova.compute.manager [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Received event network-changed-87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 693.930491] env[61991]: DEBUG nova.compute.manager [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Refreshing instance network info cache due to event network-changed-87a47717-3b2c-4160-97b7-ef566ac540c0. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 693.930784] env[61991]: DEBUG oslo_concurrency.lockutils [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] Acquiring lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.930929] env[61991]: DEBUG oslo_concurrency.lockutils [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] Acquired lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.931095] env[61991]: DEBUG nova.network.neutron [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Refreshing network info cache for port 87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.935633] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Created folder: Instances in parent group-v246790. [ 693.935930] env[61991]: DEBUG oslo.service.loopingcall [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.936699] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 693.936947] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c00228ac-eee3-400e-82ca-da8500bc933e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.957542] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de46beb-6c0e-402a-80d0-62938e83641d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.967121] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 693.967121] env[61991]: value = "task-1129409" [ 693.967121] env[61991]: _type = "Task" [ 693.967121] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.988510] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129409, 'name': CreateVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.029966] env[61991]: DEBUG nova.scheduler.client.report [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.179971] env[61991]: DEBUG nova.network.neutron [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Successfully updated port: 863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 694.385944] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.433602] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 694.434241] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-daf4f5a4-083d-46d7-9613-271edc91a17d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.443794] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 694.443794] env[61991]: value = "task-1129410" [ 694.443794] env[61991]: _type = "Task" [ 694.443794] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.451899] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129410, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.482837] env[61991]: INFO nova.compute.manager [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Rebuilding instance [ 694.486012] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129409, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.528197] env[61991]: DEBUG nova.compute.manager [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.529073] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e135bfac-1050-49cc-9166-ce92dc89d89c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.536809] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.855s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.537349] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 694.541893] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.596s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.543349] env[61991]: INFO nova.compute.claims [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.681968] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.682130] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.682281] env[61991]: DEBUG nova.network.neutron [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.902592] env[61991]: DEBUG nova.network.neutron [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updated VIF entry in instance network info cache for port 87a47717-3b2c-4160-97b7-ef566ac540c0. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 694.902952] env[61991]: DEBUG nova.network.neutron [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating instance_info_cache with network_info: [{"id": "87a47717-3b2c-4160-97b7-ef566ac540c0", "address": "fa:16:3e:e6:5f:4d", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87a47717-3b", "ovs_interfaceid": "87a47717-3b2c-4160-97b7-ef566ac540c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.911748] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.957585] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129410, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.982629] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129409, 'name': CreateVM_Task, 'duration_secs': 0.700583} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.982807] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 694.983539] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.983703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.984061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 694.984315] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f263077e-475a-4923-9785-649870970b14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.989611] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 694.989611] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521d1f4a-fe37-628c-5eb8-d2b30e98a91d" [ 694.989611] env[61991]: _type = "Task" [ 694.989611] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.998335] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521d1f4a-fe37-628c-5eb8-d2b30e98a91d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.043325] env[61991]: DEBUG nova.compute.utils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.044833] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.045019] env[61991]: DEBUG nova.network.neutron [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.047803] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 695.053268] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2eeb5c05-82bd-422c-a2f6-da2b06bd6d70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.058139] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 695.058139] env[61991]: value = "task-1129411" [ 695.058139] env[61991]: _type = "Task" [ 695.058139] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.066594] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.094758] env[61991]: DEBUG nova.policy [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5b3a828cb894e51a0f55d01fb4d133d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a861f48ec614410a472442973271b49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.220467] env[61991]: DEBUG nova.network.neutron [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.354382] env[61991]: DEBUG nova.network.neutron [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.394292] env[61991]: DEBUG nova.network.neutron [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Successfully created port: db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.405190] env[61991]: DEBUG oslo_concurrency.lockutils [req-3ae08712-2e2e-4451-9ec6-51ae90ca1cde req-3fb9bad8-d7eb-4ab5-9896-0df47bc4d057 service nova] Releasing lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.454476] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129410, 'name': CreateSnapshot_Task, 'duration_secs': 0.671243} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.454736] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 695.455486] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b8862c-70a5-4be2-8e3c-f11ed1c581f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.500089] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521d1f4a-fe37-628c-5eb8-d2b30e98a91d, 'name': SearchDatastore_Task, 'duration_secs': 0.010915} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.500394] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.500623] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 695.500846] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.500989] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.501175] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 695.501433] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9811d41-8f40-40b1-8708-7366c18ed820 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.517319] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 695.517510] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 695.518242] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9854ece9-1e1c-4287-b1a6-58e6fba93c27 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.524561] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 695.524561] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b772b-0c1d-8f4b-391d-a6ae01e2c9cd" [ 695.524561] env[61991]: _type = "Task" [ 695.524561] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.532182] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b772b-0c1d-8f4b-391d-a6ae01e2c9cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.548316] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 695.569074] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129411, 'name': PowerOffVM_Task, 'duration_secs': 0.1797} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.569074] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 695.569074] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 695.570018] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc9a2a5-49a5-4619-b9fd-eee870e19c37 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.577376] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 695.577518] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e30b3881-a421-42d8-9995-493c2a34c4fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.604640] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 695.604905] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 695.605103] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleting the datastore file [datastore2] fe78a8de-6174-47b3-9271-171c4ba76068 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 695.605680] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f06236c2-8574-477f-98dd-93d25aa2c913 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.611928] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 695.611928] env[61991]: value = "task-1129413" [ 695.611928] env[61991]: _type = "Task" [ 695.611928] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.623146] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.857581] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.857933] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Instance network_info: |[{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 695.858394] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:b6:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '863ee2c0-0c5a-49c8-a443-3c1c3d73c458', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.865862] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Creating folder: Project (518b363cb2db43c288d25c39a239c9c1). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.868767] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac2097fc-221c-4499-89c1-326bf7f7719f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.878399] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Created folder: Project (518b363cb2db43c288d25c39a239c9c1) in parent group-v246753. [ 695.879605] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Creating folder: Instances. Parent ref: group-v246794. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.879605] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e99be24-368e-4d70-b3f6-52eb83123e97 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.888832] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Created folder: Instances in parent group-v246794. [ 695.889055] env[61991]: DEBUG oslo.service.loopingcall [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.889235] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 695.889420] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbf1f969-4639-4091-9a59-1d99385f2c42 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.910604] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.910604] env[61991]: value = "task-1129416" [ 695.910604] env[61991]: _type = "Task" [ 695.910604] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.919322] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129416, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.952582] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba9f431-74af-46f6-8985-3f798aa0c2ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.962232] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5354009e-61fa-4771-958a-ddb54a09f782 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.969580] env[61991]: DEBUG nova.compute.manager [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Received event network-vif-plugged-863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.969937] env[61991]: DEBUG oslo_concurrency.lockutils [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] Acquiring lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.970295] env[61991]: DEBUG oslo_concurrency.lockutils [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.970577] env[61991]: DEBUG oslo_concurrency.lockutils [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.971036] env[61991]: DEBUG nova.compute.manager [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] No waiting events found dispatching network-vif-plugged-863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 695.971255] env[61991]: WARNING nova.compute.manager [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Received unexpected event network-vif-plugged-863ee2c0-0c5a-49c8-a443-3c1c3d73c458 for instance with vm_state building and task_state spawning. [ 695.971396] env[61991]: DEBUG nova.compute.manager [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Received event network-changed-863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.971549] env[61991]: DEBUG nova.compute.manager [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Refreshing instance network info cache due to event network-changed-863ee2c0-0c5a-49c8-a443-3c1c3d73c458. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 695.971734] env[61991]: DEBUG oslo_concurrency.lockutils [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] Acquiring lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.971867] env[61991]: DEBUG oslo_concurrency.lockutils [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] Acquired lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.972083] env[61991]: DEBUG nova.network.neutron [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Refreshing network info cache for port 863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 695.982317] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 695.983333] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-12b557a1-cf19-4d48-b090-13c03f1f5562 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.015648] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c6f121-e65c-4016-ae55-3aea3cf0f0b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.018339] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 696.018339] env[61991]: value = "task-1129417" [ 696.018339] env[61991]: _type = "Task" [ 696.018339] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.025299] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb887545-f2b0-4123-8dc0-a1eed00cf0e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.032269] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129417, 'name': CloneVM_Task} progress is 12%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.045155] env[61991]: DEBUG nova.compute.provider_tree [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.050724] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b772b-0c1d-8f4b-391d-a6ae01e2c9cd, 'name': SearchDatastore_Task, 'duration_secs': 0.032573} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.055011] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02a1f58c-9ad0-4dd9-b417-7057cacb1aab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.061990] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 696.061990] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b2ce1-ef11-072c-44f2-deda6404dab5" [ 696.061990] env[61991]: _type = "Task" [ 696.061990] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.072343] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b2ce1-ef11-072c-44f2-deda6404dab5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.121195] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24115} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.121644] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 696.121983] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 696.122243] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.420728] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129416, 'name': CreateVM_Task, 'duration_secs': 0.339439} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.421149] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 696.421578] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.421778] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.422121] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 696.422366] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cde8edd-bc08-4e88-b587-3acb977e12f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.427140] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 696.427140] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522db44d-3dc1-9ed6-9e3d-af6d370edf7c" [ 696.427140] env[61991]: _type = "Task" [ 696.427140] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.435325] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522db44d-3dc1-9ed6-9e3d-af6d370edf7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.528393] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129417, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.552477] env[61991]: DEBUG nova.scheduler.client.report [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.559041] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 696.572357] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b2ce1-ef11-072c-44f2-deda6404dab5, 'name': SearchDatastore_Task, 'duration_secs': 0.024218} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.572535] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.572755] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 9c17f844-1f0e-4f01-aadc-0f1f75a59d06/9c17f844-1f0e-4f01-aadc-0f1f75a59d06.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 696.573017] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-286da40f-ef4a-4cda-8010-c21c71d58740 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.580623] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 696.580623] env[61991]: value = "task-1129418" [ 696.580623] env[61991]: _type = "Task" [ 696.580623] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.585725] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.585949] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.586116] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.586299] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.586443] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.586593] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.586815] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.587038] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.587220] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.587381] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.587551] env[61991]: DEBUG nova.virt.hardware [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.588555] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5518ead4-f913-47a1-8cce-1f2a4bccb112 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.595390] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.598313] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e12030b-a819-4dbb-ac95-60f36f811e06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.694958] env[61991]: DEBUG nova.network.neutron [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updated VIF entry in instance network info cache for port 863ee2c0-0c5a-49c8-a443-3c1c3d73c458. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 696.695325] env[61991]: DEBUG nova.network.neutron [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.938369] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522db44d-3dc1-9ed6-9e3d-af6d370edf7c, 'name': SearchDatastore_Task, 'duration_secs': 0.011128} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.938743] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.938936] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.939166] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.939321] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.939481] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.939763] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b135aaee-9b3c-4b4f-bd13-9c97f311efa3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.971022] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.971940] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 696.972153] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd96a1d4-7151-4c34-a943-1543937bd62d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.978992] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 696.978992] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522837c0-2e0d-ed55-3136-1213fa61cfdb" [ 696.978992] env[61991]: _type = "Task" [ 696.978992] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.988273] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522837c0-2e0d-ed55-3136-1213fa61cfdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.005329] env[61991]: DEBUG nova.network.neutron [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Successfully updated port: db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 697.033066] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129417, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.059571] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.060214] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.063467] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.340s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.063783] env[61991]: DEBUG nova.objects.instance [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lazy-loading 'resources' on Instance uuid 598c0bd7-baf5-4d77-8d06-0f83a08b9685 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 697.091830] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129418, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.157187] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 697.157463] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 697.157620] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.157813] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 697.157985] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.158149] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 697.158397] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 697.158527] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 697.158694] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 697.158851] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 697.159029] env[61991]: DEBUG nova.virt.hardware [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 697.160207] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0d8705-4e66-492a-9b07-a49bd50d1b8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.168648] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa3b1df-36a9-4a54-92b2-78404c01c2d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.183723] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 697.189246] env[61991]: DEBUG oslo.service.loopingcall [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.190039] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 697.190039] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dece7a5-e97f-4b60-82c9-a59f93f49879 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.201632] env[61991]: DEBUG oslo_concurrency.lockutils [req-7ac1b1e5-a7db-49df-8285-aa0197419e10 req-22e802c0-c20e-48cf-a31c-c0484ded2b2b service nova] Releasing lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.207436] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 697.207436] env[61991]: value = "task-1129419" [ 697.207436] env[61991]: _type = "Task" [ 697.207436] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.216576] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129419, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.490442] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522837c0-2e0d-ed55-3136-1213fa61cfdb, 'name': SearchDatastore_Task, 'duration_secs': 0.085042} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.491266] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9c89861-8712-4a02-a3c9-68a2f1a069dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.496497] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 697.496497] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527234be-2156-96c5-90d5-cf491e75fca6" [ 697.496497] env[61991]: _type = "Task" [ 697.496497] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.503934] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527234be-2156-96c5-90d5-cf491e75fca6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.511642] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "refresh_cache-527561af-ba19-447f-a4ee-618824b9ce0f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.511771] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquired lock "refresh_cache-527561af-ba19-447f-a4ee-618824b9ce0f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.511959] env[61991]: DEBUG nova.network.neutron [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.530448] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129417, 'name': CloneVM_Task, 'duration_secs': 1.294711} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.530730] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Created linked-clone VM from snapshot [ 697.531497] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7031c8dd-a938-4d7f-9cf6-8a316dd1e5b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.538795] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Uploading image b07a72c1-f38c-4604-a4b1-b7e0915811f2 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 697.550622] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 697.550909] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cb7cac7b-1b6f-4d3c-b68a-4091fedd564d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.557952] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 697.557952] env[61991]: value = "task-1129420" [ 697.557952] env[61991]: _type = "Task" [ 697.557952] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.567465] env[61991]: DEBUG nova.compute.utils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.568657] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129420, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.570062] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.570062] env[61991]: DEBUG nova.network.neutron [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 697.592992] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129418, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629089} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.593132] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 9c17f844-1f0e-4f01-aadc-0f1f75a59d06/9c17f844-1f0e-4f01-aadc-0f1f75a59d06.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 697.593414] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 697.596692] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fd3e8d6-3a77-44ab-8a51-9bfff5246e77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.603281] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 697.603281] env[61991]: value = "task-1129421" [ 697.603281] env[61991]: _type = "Task" [ 697.603281] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.615381] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.621245] env[61991]: DEBUG nova.policy [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8fcf549a5154a10b9e5c1af3357e573', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c611c5aeae64333ba228bc0ffbbd03d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.718686] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129419, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.942816] env[61991]: DEBUG nova.network.neutron [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Successfully created port: 72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.966215] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d2548f-d0ea-404f-8d86-5bf53b9bebab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.973153] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bc7d98-52ef-4b40-a14e-b52f35c41cd9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.008162] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6fa6a0-1386-4f13-8fdf-8f135f3ceed7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.011538] env[61991]: DEBUG nova.compute.manager [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Received event network-vif-plugged-db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.011742] env[61991]: DEBUG oslo_concurrency.lockutils [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] Acquiring lock "527561af-ba19-447f-a4ee-618824b9ce0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.011942] env[61991]: DEBUG oslo_concurrency.lockutils [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] Lock "527561af-ba19-447f-a4ee-618824b9ce0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.012124] env[61991]: DEBUG oslo_concurrency.lockutils [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] Lock "527561af-ba19-447f-a4ee-618824b9ce0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.012290] env[61991]: DEBUG nova.compute.manager [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] No waiting events found dispatching network-vif-plugged-db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 698.012449] env[61991]: WARNING nova.compute.manager [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Received unexpected event network-vif-plugged-db700598-7ac7-4c06-9882-3206f0ae86ad for instance with vm_state building and task_state spawning. [ 698.012609] env[61991]: DEBUG nova.compute.manager [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Received event network-changed-db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.012749] env[61991]: DEBUG nova.compute.manager [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Refreshing instance network info cache due to event network-changed-db700598-7ac7-4c06-9882-3206f0ae86ad. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 698.012912] env[61991]: DEBUG oslo_concurrency.lockutils [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] Acquiring lock "refresh_cache-527561af-ba19-447f-a4ee-618824b9ce0f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.027296] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed72ee94-44e6-4330-aeec-a30902166246 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.028603] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527234be-2156-96c5-90d5-cf491e75fca6, 'name': SearchDatastore_Task, 'duration_secs': 0.016356} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.029172] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.029463] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 698.029991] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-998b9952-2bd9-4ade-98b5-b838636d81d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.040929] env[61991]: DEBUG nova.compute.provider_tree [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.047213] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 698.047213] env[61991]: value = "task-1129422" [ 698.047213] env[61991]: _type = "Task" [ 698.047213] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.055935] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129422, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.068276] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129420, 'name': Destroy_Task, 'duration_secs': 0.457022} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.068993] env[61991]: DEBUG nova.network.neutron [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.070744] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Destroyed the VM [ 698.071108] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 698.071564] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a81ac489-514e-4374-8b53-c85c0094083e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.077396] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.078326] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 698.078326] env[61991]: value = "task-1129423" [ 698.078326] env[61991]: _type = "Task" [ 698.078326] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.086064] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129423, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.112230] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085108} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.112485] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.113198] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3498cf-c09f-4a3a-be01-efcd0c9ca4c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.135501] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 9c17f844-1f0e-4f01-aadc-0f1f75a59d06/9c17f844-1f0e-4f01-aadc-0f1f75a59d06.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.141018] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d85b6c45-37ee-4e22-9e3a-ea71448bb947 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.158248] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 698.158248] env[61991]: value = "task-1129424" [ 698.158248] env[61991]: _type = "Task" [ 698.158248] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.168221] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129424, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.224451] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129419, 'name': CreateVM_Task, 'duration_secs': 0.538145} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.224451] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 698.224451] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.224451] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.224663] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 698.224829] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562a2e34-b1d6-4c08-9dfa-0d915a7b9b6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.232251] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 698.232251] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d4854e-1e17-e261-5e8b-f2433ecffd78" [ 698.232251] env[61991]: _type = "Task" [ 698.232251] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.241560] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d4854e-1e17-e261-5e8b-f2433ecffd78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.365163] env[61991]: DEBUG nova.network.neutron [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Updating instance_info_cache with network_info: [{"id": "db700598-7ac7-4c06-9882-3206f0ae86ad", "address": "fa:16:3e:62:58:1b", "network": {"id": "661540b9-03e6-4a28-954d-0ff53122c3d6", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1277905508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a861f48ec614410a472442973271b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb700598-7a", "ovs_interfaceid": "db700598-7ac7-4c06-9882-3206f0ae86ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.544151] env[61991]: DEBUG nova.scheduler.client.report [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.557010] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129422, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497633} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.557298] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 698.557504] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.557755] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bdc82e8f-aaaf-4716-80c8-0cdfcbf63e2c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.566401] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 698.566401] env[61991]: value = "task-1129425" [ 698.566401] env[61991]: _type = "Task" [ 698.566401] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.577257] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.594601] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129423, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.670684] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129424, 'name': ReconfigVM_Task, 'duration_secs': 0.335745} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.670979] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 9c17f844-1f0e-4f01-aadc-0f1f75a59d06/9c17f844-1f0e-4f01-aadc-0f1f75a59d06.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 698.671681] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b885b511-90c9-4e19-b1e9-f344e290bb23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.679424] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 698.679424] env[61991]: value = "task-1129426" [ 698.679424] env[61991]: _type = "Task" [ 698.679424] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.690970] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129426, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.742660] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d4854e-1e17-e261-5e8b-f2433ecffd78, 'name': SearchDatastore_Task, 'duration_secs': 0.059858} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.742991] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.743249] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.743495] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.743642] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.743855] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.744430] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31682973-9184-43b8-a840-ce819e17c05f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.756685] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.756890] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.757941] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e75c637-0038-4432-a844-2b9e9869cdb1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.763400] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 698.763400] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522debde-2e0d-4642-66dd-a75a0ed22aa8" [ 698.763400] env[61991]: _type = "Task" [ 698.763400] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.771866] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522debde-2e0d-4642-66dd-a75a0ed22aa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.872665] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Releasing lock "refresh_cache-527561af-ba19-447f-a4ee-618824b9ce0f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.873040] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Instance network_info: |[{"id": "db700598-7ac7-4c06-9882-3206f0ae86ad", "address": "fa:16:3e:62:58:1b", "network": {"id": "661540b9-03e6-4a28-954d-0ff53122c3d6", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1277905508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a861f48ec614410a472442973271b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb700598-7a", "ovs_interfaceid": "db700598-7ac7-4c06-9882-3206f0ae86ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 698.873356] env[61991]: DEBUG oslo_concurrency.lockutils [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] Acquired lock "refresh_cache-527561af-ba19-447f-a4ee-618824b9ce0f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.873533] env[61991]: DEBUG nova.network.neutron [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Refreshing network info cache for port db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 698.875148] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:58:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db700598-7ac7-4c06-9882-3206f0ae86ad', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 698.882935] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Creating folder: Project (7a861f48ec614410a472442973271b49). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 698.884158] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-283f7460-848a-48bc-b150-6512b70ae8f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.898306] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Created folder: Project (7a861f48ec614410a472442973271b49) in parent group-v246753. [ 698.898550] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Creating folder: Instances. Parent ref: group-v246799. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 698.898765] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8fea307b-9df7-41d8-a739-bb0b410a14db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.909282] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Created folder: Instances in parent group-v246799. [ 698.909539] env[61991]: DEBUG oslo.service.loopingcall [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.909735] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 698.910033] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74cedda3-6763-4abf-9ae8-93ad5cc34dba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.929448] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 698.929448] env[61991]: value = "task-1129429" [ 698.929448] env[61991]: _type = "Task" [ 698.929448] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.937958] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129429, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.049085] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.985s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.052175] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.976s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.052175] env[61991]: DEBUG nova.objects.instance [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lazy-loading 'resources' on Instance uuid b1c1c4e4-a82b-4066-8ee8-342d5ad18d29 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 699.077621] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.217561} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.078757] env[61991]: INFO nova.scheduler.client.report [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted allocations for instance 598c0bd7-baf5-4d77-8d06-0f83a08b9685 [ 699.079763] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 699.081402] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e137d18-1b88-4ccc-acb3-e3a933a3edcf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.092594] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.115685] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 699.119244] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95853e4e-9748-4b70-af62-720e88e06ec2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.134687] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129423, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.142542] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 699.142542] env[61991]: value = "task-1129430" [ 699.142542] env[61991]: _type = "Task" [ 699.142542] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.144827] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.145056] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.145225] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.145414] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.145557] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.145700] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.145909] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.146083] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.146252] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.146412] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.146580] env[61991]: DEBUG nova.virt.hardware [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.147819] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c68936-79df-46e2-950b-7d21cd0ef1bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.162992] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fcc895-b454-4e01-80bf-5e5ff0920307 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.169025] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.192169] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129426, 'name': Rename_Task, 'duration_secs': 0.461839} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.192476] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 699.192722] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e714e10-13a7-44ca-a0f0-cada7493f532 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.199726] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 699.199726] env[61991]: value = "task-1129431" [ 699.199726] env[61991]: _type = "Task" [ 699.199726] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.208224] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129431, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.273725] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522debde-2e0d-4642-66dd-a75a0ed22aa8, 'name': SearchDatastore_Task, 'duration_secs': 0.020449} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.274669] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f13e6a28-069f-4998-9b23-a966d073e333 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.280683] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 699.280683] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525c2016-8eab-987d-f031-4e40d58ca35a" [ 699.280683] env[61991]: _type = "Task" [ 699.280683] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.289171] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525c2016-8eab-987d-f031-4e40d58ca35a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.440763] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129429, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.595874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ad07c26-dd1d-45b6-9f06-4ff84178964d tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "598c0bd7-baf5-4d77-8d06-0f83a08b9685" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.804s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.610891] env[61991]: DEBUG oslo_vmware.api [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129423, 'name': RemoveSnapshot_Task, 'duration_secs': 1.245886} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.611229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 699.657720] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.708250] env[61991]: DEBUG nova.network.neutron [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Updated VIF entry in instance network info cache for port db700598-7ac7-4c06-9882-3206f0ae86ad. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 699.708250] env[61991]: DEBUG nova.network.neutron [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Updating instance_info_cache with network_info: [{"id": "db700598-7ac7-4c06-9882-3206f0ae86ad", "address": "fa:16:3e:62:58:1b", "network": {"id": "661540b9-03e6-4a28-954d-0ff53122c3d6", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1277905508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a861f48ec614410a472442973271b49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb700598-7a", "ovs_interfaceid": "db700598-7ac7-4c06-9882-3206f0ae86ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.714474] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129431, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.730929] env[61991]: DEBUG nova.network.neutron [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Successfully updated port: 72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 699.800325] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525c2016-8eab-987d-f031-4e40d58ca35a, 'name': SearchDatastore_Task, 'duration_secs': 0.033043} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.803231] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.803513] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.804153] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1f05e12-381c-4187-b845-233d00584d88 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.811225] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 699.811225] env[61991]: value = "task-1129432" [ 699.811225] env[61991]: _type = "Task" [ 699.811225] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.820601] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129432, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.944431] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129429, 'name': CreateVM_Task, 'duration_secs': 0.887173} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.944656] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 699.945526] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.945707] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.946055] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 699.946425] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f47306b-c3f7-4e72-a1b8-fbbbed057870 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.953743] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 699.953743] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52524c18-ea2e-7251-a1a2-f7bd990619b4" [ 699.953743] env[61991]: _type = "Task" [ 699.953743] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.964084] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52524c18-ea2e-7251-a1a2-f7bd990619b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.007219] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803eba2c-b9df-4631-94bc-58c32f417fcd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.013703] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43231434-3a83-4441-925a-c4258bca782a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.047474] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c84c30e-c700-4f2b-96cf-31c11ddfaed6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.055450] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d3e122-c2e6-4659-893e-fba28865b0f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.069538] env[61991]: DEBUG nova.compute.provider_tree [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.109789] env[61991]: DEBUG nova.compute.manager [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Received event network-vif-plugged-72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.109984] env[61991]: DEBUG oslo_concurrency.lockutils [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] Acquiring lock "43cb61e7-1748-40d9-a287-1179c8219c2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.110217] env[61991]: DEBUG oslo_concurrency.lockutils [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.110401] env[61991]: DEBUG oslo_concurrency.lockutils [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.110562] env[61991]: DEBUG nova.compute.manager [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] No waiting events found dispatching network-vif-plugged-72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 700.110726] env[61991]: WARNING nova.compute.manager [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Received unexpected event network-vif-plugged-72dca87d-ec73-4972-8883-0f9dd0284313 for instance with vm_state building and task_state spawning. [ 700.111153] env[61991]: DEBUG nova.compute.manager [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Received event network-changed-72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.111333] env[61991]: DEBUG nova.compute.manager [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Refreshing instance network info cache due to event network-changed-72dca87d-ec73-4972-8883-0f9dd0284313. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 700.111519] env[61991]: DEBUG oslo_concurrency.lockutils [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] Acquiring lock "refresh_cache-43cb61e7-1748-40d9-a287-1179c8219c2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.111689] env[61991]: DEBUG oslo_concurrency.lockutils [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] Acquired lock "refresh_cache-43cb61e7-1748-40d9-a287-1179c8219c2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.111842] env[61991]: DEBUG nova.network.neutron [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Refreshing network info cache for port 72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 700.120637] env[61991]: WARNING nova.compute.manager [None req-10034dd3-cecf-44ee-9584-735a34501efc tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Image not found during snapshot: nova.exception.ImageNotFound: Image b07a72c1-f38c-4604-a4b1-b7e0915811f2 could not be found. [ 700.160526] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129430, 'name': ReconfigVM_Task, 'duration_secs': 0.673934} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.160905] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfigured VM instance instance-0000002b to attach disk [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 700.161597] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-420f577a-ad38-4955-a5d6-4f7f1117f52c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.170028] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 700.170028] env[61991]: value = "task-1129433" [ 700.170028] env[61991]: _type = "Task" [ 700.170028] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.180021] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129433, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.209907] env[61991]: DEBUG oslo_concurrency.lockutils [req-635eb224-ca0e-4dbd-8c93-8ab5e4c67087 req-24811b86-8d45-4c53-bb50-fbb353e39369 service nova] Releasing lock "refresh_cache-527561af-ba19-447f-a4ee-618824b9ce0f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.213309] env[61991]: DEBUG oslo_vmware.api [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129431, 'name': PowerOnVM_Task, 'duration_secs': 0.760815} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.213559] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 700.213808] env[61991]: INFO nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Took 8.96 seconds to spawn the instance on the hypervisor. [ 700.214023] env[61991]: DEBUG nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 700.214834] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef4804b-0667-4c03-b875-485b3444623b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.238027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "refresh_cache-43cb61e7-1748-40d9-a287-1179c8219c2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.324872] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129432, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.464180] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52524c18-ea2e-7251-a1a2-f7bd990619b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010692} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.464180] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.464366] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.464470] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.464613] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.464785] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.465047] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92f08f8a-0ca9-4732-9285-8d67088d18b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.472705] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.472880] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 700.473574] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3da2ad8e-0371-46da-923a-27ed4ee161dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.478438] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 700.478438] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52321300-0da3-c73c-8adb-9b1e72b36a84" [ 700.478438] env[61991]: _type = "Task" [ 700.478438] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.485726] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52321300-0da3-c73c-8adb-9b1e72b36a84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.573109] env[61991]: DEBUG nova.scheduler.client.report [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.659468] env[61991]: DEBUG nova.network.neutron [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.680165] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129433, 'name': Rename_Task, 'duration_secs': 0.209197} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.680357] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 700.680590] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d40ccb7d-5ffd-4f0c-ae93-564d22cc5f15 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.687065] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 700.687065] env[61991]: value = "task-1129434" [ 700.687065] env[61991]: _type = "Task" [ 700.687065] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.694159] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.734025] env[61991]: INFO nova.compute.manager [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Took 33.03 seconds to build instance. [ 700.780756] env[61991]: DEBUG nova.network.neutron [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.821217] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129432, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573593} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.821481] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.821758] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.822029] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da1841c5-3f0b-46ce-8096-2bf102c7aa20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.828160] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 700.828160] env[61991]: value = "task-1129435" [ 700.828160] env[61991]: _type = "Task" [ 700.828160] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.835817] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129435, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.989388] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52321300-0da3-c73c-8adb-9b1e72b36a84, 'name': SearchDatastore_Task, 'duration_secs': 0.00749} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.990316] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-535aa187-1e2c-4a1d-ab4f-29c65857364b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.996119] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 700.996119] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52418aaa-6802-6930-d0fc-a2be5c60e02e" [ 700.996119] env[61991]: _type = "Task" [ 700.996119] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.003419] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52418aaa-6802-6930-d0fc-a2be5c60e02e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.082020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.082020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.640s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.083364] env[61991]: INFO nova.compute.claims [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.118531] env[61991]: INFO nova.scheduler.client.report [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted allocations for instance b1c1c4e4-a82b-4066-8ee8-342d5ad18d29 [ 701.197474] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129434, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.236753] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d46ab0a-fcf0-446e-ab9b-5725f2b48188 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.989s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.283870] env[61991]: DEBUG oslo_concurrency.lockutils [req-b75f3c3c-2bcc-451a-9e9d-9a12d709c10a req-212d88d3-2210-4bf1-99f8-01c7913b1756 service nova] Releasing lock "refresh_cache-43cb61e7-1748-40d9-a287-1179c8219c2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.284604] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquired lock "refresh_cache-43cb61e7-1748-40d9-a287-1179c8219c2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.284845] env[61991]: DEBUG nova.network.neutron [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 701.338840] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129435, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.506943] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52418aaa-6802-6930-d0fc-a2be5c60e02e, 'name': SearchDatastore_Task, 'duration_secs': 0.009622} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.507236] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.507495] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 527561af-ba19-447f-a4ee-618824b9ce0f/527561af-ba19-447f-a4ee-618824b9ce0f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 701.507794] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4a45193-1149-45e6-ab7d-7e36c7514889 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.513779] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 701.513779] env[61991]: value = "task-1129436" [ 701.513779] env[61991]: _type = "Task" [ 701.513779] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.522691] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.624978] env[61991]: DEBUG oslo_concurrency.lockutils [None req-64bac606-f047-4508-9b19-2f83a535326f tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "b1c1c4e4-a82b-4066-8ee8-342d5ad18d29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.668s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.702652] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "812d3491-0598-4171-b905-5304ebf6ae37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.702937] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "812d3491-0598-4171-b905-5304ebf6ae37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.702937] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "812d3491-0598-4171-b905-5304ebf6ae37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.703187] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "812d3491-0598-4171-b905-5304ebf6ae37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.703334] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "812d3491-0598-4171-b905-5304ebf6ae37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.705622] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129434, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.705789] env[61991]: INFO nova.compute.manager [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Terminating instance [ 701.709552] env[61991]: DEBUG nova.compute.manager [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.709659] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 701.710541] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf15896-04ca-4637-9cf2-a5cb46d33703 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.722898] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 701.723271] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fe0962a-1a80-4d63-bf97-e83028c61f6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.733071] env[61991]: DEBUG oslo_vmware.api [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 701.733071] env[61991]: value = "task-1129437" [ 701.733071] env[61991]: _type = "Task" [ 701.733071] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.739252] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.749554] env[61991]: DEBUG oslo_vmware.api [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.842664] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129435, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.523257} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.842664] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.845861] env[61991]: DEBUG nova.network.neutron [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.850595] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e8f791-935e-4c15-b178-197daae04ea5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.875277] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.876017] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c38d986-3698-45cb-8148-5cf6742e4040 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.904594] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 701.904594] env[61991]: value = "task-1129438" [ 701.904594] env[61991]: _type = "Task" [ 701.904594] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.916071] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129438, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.027803] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129436, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465213} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.028152] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 527561af-ba19-447f-a4ee-618824b9ce0f/527561af-ba19-447f-a4ee-618824b9ce0f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 702.028391] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 702.028681] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eaf0f544-c10a-4784-be63-710941d0f462 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.036119] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 702.036119] env[61991]: value = "task-1129439" [ 702.036119] env[61991]: _type = "Task" [ 702.036119] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.045733] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129439, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.068990] env[61991]: DEBUG nova.network.neutron [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Updating instance_info_cache with network_info: [{"id": "72dca87d-ec73-4972-8883-0f9dd0284313", "address": "fa:16:3e:94:50:11", "network": {"id": "e4a8bac4-0038-4c09-8c3a-d38255a24196", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-197401537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c611c5aeae64333ba228bc0ffbbd03d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72dca87d-ec", "ovs_interfaceid": "72dca87d-ec73-4972-8883-0f9dd0284313", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.139996] env[61991]: DEBUG nova.compute.manager [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Received event network-changed-87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.139996] env[61991]: DEBUG nova.compute.manager [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Refreshing instance network info cache due to event network-changed-87a47717-3b2c-4160-97b7-ef566ac540c0. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.140424] env[61991]: DEBUG oslo_concurrency.lockutils [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] Acquiring lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.140568] env[61991]: DEBUG oslo_concurrency.lockutils [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] Acquired lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.140724] env[61991]: DEBUG nova.network.neutron [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Refreshing network info cache for port 87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 702.202057] env[61991]: DEBUG oslo_vmware.api [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129434, 'name': PowerOnVM_Task, 'duration_secs': 1.198338} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.202338] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 702.202536] env[61991]: INFO nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Took 8.32 seconds to spawn the instance on the hypervisor. [ 702.202710] env[61991]: DEBUG nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 702.203495] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e70964-e281-412f-9e51-df62412b42ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.245124] env[61991]: DEBUG oslo_vmware.api [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129437, 'name': PowerOffVM_Task, 'duration_secs': 0.318836} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.245442] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 702.245643] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 702.245917] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb2461d0-0b12-4a1e-bfe4-e56a12258b61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.268126] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.304641] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 702.304902] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 702.305155] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleting the datastore file [datastore2] 812d3491-0598-4171-b905-5304ebf6ae37 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 702.307679] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5f8acf7-6648-466b-80f2-18a734d2c55a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.314713] env[61991]: DEBUG oslo_vmware.api [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 702.314713] env[61991]: value = "task-1129441" [ 702.314713] env[61991]: _type = "Task" [ 702.314713] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.323877] env[61991]: DEBUG oslo_vmware.api [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129441, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.419631] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129438, 'name': ReconfigVM_Task, 'duration_secs': 0.283558} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.420288] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Reconfigured VM instance instance-00000029 to attach disk [datastore1] fe78a8de-6174-47b3-9271-171c4ba76068/fe78a8de-6174-47b3-9271-171c4ba76068.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.420514] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-847edc58-6590-43c5-ba01-69b0b4033864 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.426434] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3741e4d-38e0-4b5d-9e3a-19f380eb27ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.432015] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 702.432015] env[61991]: value = "task-1129442" [ 702.432015] env[61991]: _type = "Task" [ 702.432015] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.435911] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71662cff-2cb5-4931-afc4-903a5716b3da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.441558] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129442, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.470670] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600bbaa7-6712-48e5-9724-ef34e322a54a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.478301] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960991fd-6fc5-4be0-8d9d-6c69d5a9c3bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.492245] env[61991]: DEBUG nova.compute.provider_tree [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.549697] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129439, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070089} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.549697] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 702.549697] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b289160-2599-4bb3-bcbb-8ef0263998a2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.573026] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 527561af-ba19-447f-a4ee-618824b9ce0f/527561af-ba19-447f-a4ee-618824b9ce0f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 702.573026] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faf5dc82-ae40-48d9-be4c-67dcbe931374 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.587242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Releasing lock "refresh_cache-43cb61e7-1748-40d9-a287-1179c8219c2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.587564] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Instance network_info: |[{"id": "72dca87d-ec73-4972-8883-0f9dd0284313", "address": "fa:16:3e:94:50:11", "network": {"id": "e4a8bac4-0038-4c09-8c3a-d38255a24196", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-197401537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5c611c5aeae64333ba228bc0ffbbd03d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72dca87d-ec", "ovs_interfaceid": "72dca87d-ec73-4972-8883-0f9dd0284313", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 702.587977] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:50:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72dca87d-ec73-4972-8883-0f9dd0284313', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 702.595245] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Creating folder: Project (5c611c5aeae64333ba228bc0ffbbd03d). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 702.595972] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1eb4a635-6cbe-412d-8d61-02c35da2e6cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.600989] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 702.600989] env[61991]: value = "task-1129443" [ 702.600989] env[61991]: _type = "Task" [ 702.600989] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.606334] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Created folder: Project (5c611c5aeae64333ba228bc0ffbbd03d) in parent group-v246753. [ 702.606458] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Creating folder: Instances. Parent ref: group-v246802. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 702.609669] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2576891-6db1-4616-aad2-674689b70eee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.611244] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129443, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.621074] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Created folder: Instances in parent group-v246802. [ 702.621455] env[61991]: DEBUG oslo.service.loopingcall [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.621750] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 702.622060] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d87bae6-d5ec-4e5d-94d1-9dba29313d35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.643033] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 702.643033] env[61991]: value = "task-1129446" [ 702.643033] env[61991]: _type = "Task" [ 702.643033] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.654301] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129446, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.724789] env[61991]: INFO nova.compute.manager [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Took 32.91 seconds to build instance. [ 702.825827] env[61991]: DEBUG oslo_vmware.api [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129441, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151498} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.825827] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 702.826087] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 702.826087] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 702.826217] env[61991]: INFO nova.compute.manager [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Took 1.12 seconds to destroy the instance on the hypervisor. [ 702.826455] env[61991]: DEBUG oslo.service.loopingcall [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.826635] env[61991]: DEBUG nova.compute.manager [-] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.826726] env[61991]: DEBUG nova.network.neutron [-] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 702.942430] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129442, 'name': Rename_Task, 'duration_secs': 0.153758} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.942529] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.942689] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9325b3d8-371e-4e59-b672-359c171d2c47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.948953] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 702.948953] env[61991]: value = "task-1129447" [ 702.948953] env[61991]: _type = "Task" [ 702.948953] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.960916] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129447, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.995199] env[61991]: DEBUG nova.scheduler.client.report [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.034628] env[61991]: DEBUG nova.network.neutron [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updated VIF entry in instance network info cache for port 87a47717-3b2c-4160-97b7-ef566ac540c0. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 703.035080] env[61991]: DEBUG nova.network.neutron [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating instance_info_cache with network_info: [{"id": "87a47717-3b2c-4160-97b7-ef566ac540c0", "address": "fa:16:3e:e6:5f:4d", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87a47717-3b", "ovs_interfaceid": "87a47717-3b2c-4160-97b7-ef566ac540c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.117358] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129443, 'name': ReconfigVM_Task, 'duration_secs': 0.272564} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.117358] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 527561af-ba19-447f-a4ee-618824b9ce0f/527561af-ba19-447f-a4ee-618824b9ce0f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 703.117358] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbe8700b-2830-49a7-8efe-5be1419d68b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.123721] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 703.123721] env[61991]: value = "task-1129448" [ 703.123721] env[61991]: _type = "Task" [ 703.123721] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.135074] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129448, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.152957] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129446, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.229977] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4d07c097-9ed6-46ba-84f2-65ec81ab2ba4 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.990s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.234509] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.234836] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.266182] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.266414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.306252] env[61991]: DEBUG nova.compute.manager [req-88993ec4-1f5c-4169-8e37-3e6fa2c0924d req-773cb05c-623f-4efe-b983-f8c7365c90e6 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Received event network-vif-deleted-6d378d4e-f7a2-47e8-a0f1-135318bd5381 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.306486] env[61991]: INFO nova.compute.manager [req-88993ec4-1f5c-4169-8e37-3e6fa2c0924d req-773cb05c-623f-4efe-b983-f8c7365c90e6 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Neutron deleted interface 6d378d4e-f7a2-47e8-a0f1-135318bd5381; detaching it from the instance and deleting it from the info cache [ 703.307028] env[61991]: DEBUG nova.network.neutron [req-88993ec4-1f5c-4169-8e37-3e6fa2c0924d req-773cb05c-623f-4efe-b983-f8c7365c90e6 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.460348] env[61991]: DEBUG oslo_vmware.api [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129447, 'name': PowerOnVM_Task, 'duration_secs': 0.463179} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.460348] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 703.460348] env[61991]: DEBUG nova.compute.manager [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 703.460586] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322d5ebb-8fb2-42e4-96b6-1b0805b22d5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.499841] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.500457] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.503333] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.690s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.505138] env[61991]: INFO nova.compute.claims [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.540669] env[61991]: DEBUG oslo_concurrency.lockutils [req-b4547685-6425-47c7-bbf6-442640a5debc req-29096b9a-1948-4265-96c9-4359e76c0ac9 service nova] Releasing lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.633923] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129448, 'name': Rename_Task, 'duration_secs': 0.162454} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.634235] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 703.634481] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5562c8a6-5986-4ad4-8f53-173c3be970ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.640440] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 703.640440] env[61991]: value = "task-1129449" [ 703.640440] env[61991]: _type = "Task" [ 703.640440] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.650799] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.655489] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129446, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.674459] env[61991]: DEBUG nova.network.neutron [-] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.736864] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.811353] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b744565f-9819-4fea-bb6c-298c8c32bf76 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.827947] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1660ff90-3ef8-4a7e-8a3e-879e4c54de1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.860324] env[61991]: DEBUG nova.compute.manager [req-88993ec4-1f5c-4169-8e37-3e6fa2c0924d req-773cb05c-623f-4efe-b983-f8c7365c90e6 service nova] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Detach interface failed, port_id=6d378d4e-f7a2-47e8-a0f1-135318bd5381, reason: Instance 812d3491-0598-4171-b905-5304ebf6ae37 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 703.977627] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.013624] env[61991]: DEBUG nova.compute.utils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.016025] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.016025] env[61991]: DEBUG nova.network.neutron [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.072479] env[61991]: DEBUG nova.policy [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a385cf261940a19eb17046958920ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3015571d285418aae0fca50b563394e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.152695] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129449, 'name': PowerOnVM_Task} progress is 1%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.155959] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129446, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.178137] env[61991]: INFO nova.compute.manager [-] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Took 1.35 seconds to deallocate network for instance. [ 704.259381] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.438672] env[61991]: DEBUG nova.network.neutron [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Successfully created port: 91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.519509] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.660593] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129449, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.664369] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129446, 'name': CreateVM_Task, 'duration_secs': 1.617056} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.666970] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 704.667947] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.668150] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.668782] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 704.669266] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a54bf475-9747-4078-ba16-48a060f989ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.674417] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 704.674417] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a534b-9a3d-9d1d-53d0-94951fc73dee" [ 704.674417] env[61991]: _type = "Task" [ 704.674417] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.685827] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.692417] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a534b-9a3d-9d1d-53d0-94951fc73dee, 'name': SearchDatastore_Task, 'duration_secs': 0.010588} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.692848] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.693132] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.693391] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.693593] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.693798] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.694133] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef112ac9-e5ff-4b4e-ad0c-1c6d8b12d13c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.707624] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.707826] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 704.708736] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a199139-831f-404e-8d32-a5a1447d0656 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.722569] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 704.722569] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523ac67f-7ccf-3d82-8971-fddfcf9fe228" [ 704.722569] env[61991]: _type = "Task" [ 704.722569] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.736287] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523ac67f-7ccf-3d82-8971-fddfcf9fe228, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.928754] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ceea5f3-af89-46af-a6c7-a620a1639d70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.936916] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a32644e-7a61-4255-9987-02ce0c787248 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.966594] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d542e7-2793-4a34-b278-fece59043486 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.974397] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60712d94-6ee1-4739-ad1f-700df2c95d4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.988823] env[61991]: DEBUG nova.compute.provider_tree [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.120748] env[61991]: DEBUG nova.compute.manager [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 705.151673] env[61991]: DEBUG oslo_vmware.api [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129449, 'name': PowerOnVM_Task, 'duration_secs': 1.187071} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.151946] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 705.152161] env[61991]: INFO nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Took 8.59 seconds to spawn the instance on the hypervisor. [ 705.152349] env[61991]: DEBUG nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 705.153118] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27e9f5e-2815-47e6-886d-49c77e1ae116 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.221030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "fe78a8de-6174-47b3-9271-171c4ba76068" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.221030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "fe78a8de-6174-47b3-9271-171c4ba76068" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.221030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "fe78a8de-6174-47b3-9271-171c4ba76068-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.221195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "fe78a8de-6174-47b3-9271-171c4ba76068-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.221335] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "fe78a8de-6174-47b3-9271-171c4ba76068-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.226891] env[61991]: INFO nova.compute.manager [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Terminating instance [ 705.228340] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "refresh_cache-fe78a8de-6174-47b3-9271-171c4ba76068" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.228421] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "refresh_cache-fe78a8de-6174-47b3-9271-171c4ba76068" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.228594] env[61991]: DEBUG nova.network.neutron [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.235749] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523ac67f-7ccf-3d82-8971-fddfcf9fe228, 'name': SearchDatastore_Task, 'duration_secs': 0.014972} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.237170] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34bc0b78-4584-4c26-a4d4-864ee58455f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.243275] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 705.243275] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e546f4-1e0d-b865-c009-e1a7bbc1de61" [ 705.243275] env[61991]: _type = "Task" [ 705.243275] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.251287] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e546f4-1e0d-b865-c009-e1a7bbc1de61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.491563] env[61991]: DEBUG nova.scheduler.client.report [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.528199] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.553310] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.553600] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.553794] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.554043] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.554233] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.554397] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.554637] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.555547] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.555547] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.555547] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.555547] env[61991]: DEBUG nova.virt.hardware [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.557371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab56849-5bee-4509-8da7-5fabe81eedac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.565538] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709a66cc-9b9a-42d1-a813-8d0906a2a4e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.643076] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.670918] env[61991]: INFO nova.compute.manager [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Took 31.27 seconds to build instance. [ 705.747740] env[61991]: DEBUG nova.network.neutron [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.756800] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e546f4-1e0d-b865-c009-e1a7bbc1de61, 'name': SearchDatastore_Task, 'duration_secs': 0.032295} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.757058] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.757313] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 43cb61e7-1748-40d9-a287-1179c8219c2a/43cb61e7-1748-40d9-a287-1179c8219c2a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.757561] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33a7b174-9483-49e1-9a99-831974965b33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.764150] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 705.764150] env[61991]: value = "task-1129450" [ 705.764150] env[61991]: _type = "Task" [ 705.764150] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.771769] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129450, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.809921] env[61991]: DEBUG nova.network.neutron [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.992759] env[61991]: DEBUG nova.compute.manager [req-1111a14d-598f-4de9-8442-81dab2e9bee2 req-0efa606b-ea37-448a-9158-20bdd9799544 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Received event network-vif-plugged-91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.993228] env[61991]: DEBUG oslo_concurrency.lockutils [req-1111a14d-598f-4de9-8442-81dab2e9bee2 req-0efa606b-ea37-448a-9158-20bdd9799544 service nova] Acquiring lock "1715711a-72c5-4e86-88e0-cecf922cb42f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.993332] env[61991]: DEBUG oslo_concurrency.lockutils [req-1111a14d-598f-4de9-8442-81dab2e9bee2 req-0efa606b-ea37-448a-9158-20bdd9799544 service nova] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.993471] env[61991]: DEBUG oslo_concurrency.lockutils [req-1111a14d-598f-4de9-8442-81dab2e9bee2 req-0efa606b-ea37-448a-9158-20bdd9799544 service nova] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.993641] env[61991]: DEBUG nova.compute.manager [req-1111a14d-598f-4de9-8442-81dab2e9bee2 req-0efa606b-ea37-448a-9158-20bdd9799544 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] No waiting events found dispatching network-vif-plugged-91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 705.993974] env[61991]: WARNING nova.compute.manager [req-1111a14d-598f-4de9-8442-81dab2e9bee2 req-0efa606b-ea37-448a-9158-20bdd9799544 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Received unexpected event network-vif-plugged-91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f for instance with vm_state building and task_state spawning. [ 705.996048] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.996831] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.999841] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.979s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.003056] env[61991]: DEBUG nova.objects.instance [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lazy-loading 'resources' on Instance uuid a2606541-ff39-4d03-bd19-d0ee8e035c08 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 706.085689] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "527561af-ba19-447f-a4ee-618824b9ce0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.176061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8244229-c71d-4360-8e8b-582c9419662c tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "527561af-ba19-447f-a4ee-618824b9ce0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.392s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.176061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "527561af-ba19-447f-a4ee-618824b9ce0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.089s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.176061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "527561af-ba19-447f-a4ee-618824b9ce0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.176061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "527561af-ba19-447f-a4ee-618824b9ce0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.176498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "527561af-ba19-447f-a4ee-618824b9ce0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.182382] env[61991]: INFO nova.compute.manager [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Terminating instance [ 706.185453] env[61991]: DEBUG nova.compute.manager [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.186034] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.187586] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b35b98-e8cf-440f-b798-1205da5f4ab2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.198375] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 706.199066] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b29ab826-62b9-4922-9340-f38e3ef53f5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.207792] env[61991]: DEBUG oslo_vmware.api [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 706.207792] env[61991]: value = "task-1129451" [ 706.207792] env[61991]: _type = "Task" [ 706.207792] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.219885] env[61991]: DEBUG oslo_vmware.api [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.277912] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129450, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.315578] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "refresh_cache-fe78a8de-6174-47b3-9271-171c4ba76068" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.315578] env[61991]: DEBUG nova.compute.manager [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.315578] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.315578] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167dd931-ea8a-44db-b21c-560c7b1cdfc0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.324219] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 706.324607] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a68adbbf-89ac-42ca-92ce-c8bc0b53c81e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.335474] env[61991]: DEBUG oslo_vmware.api [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 706.335474] env[61991]: value = "task-1129452" [ 706.335474] env[61991]: _type = "Task" [ 706.335474] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.344608] env[61991]: DEBUG oslo_vmware.api [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.346171] env[61991]: DEBUG nova.network.neutron [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Successfully updated port: 91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.503340] env[61991]: DEBUG nova.compute.utils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.506166] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.506474] env[61991]: DEBUG nova.network.neutron [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 706.560126] env[61991]: DEBUG nova.policy [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b5d90a913fc41e282e050bb9af628d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adbd4ea9b0744f28a0d4a46b4a04d683', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.682549] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.723765] env[61991]: DEBUG oslo_vmware.api [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129451, 'name': PowerOffVM_Task, 'duration_secs': 0.208221} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.724104] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 706.724240] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 706.724483] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-622fd375-443f-413f-a44a-6abde99b3925 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.774656] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129450, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553983} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.777199] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 43cb61e7-1748-40d9-a287-1179c8219c2a/43cb61e7-1748-40d9-a287-1179c8219c2a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.777199] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.777675] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea3fb7ae-22a3-4cf1-a48e-df04ec08be7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.784035] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 706.784035] env[61991]: value = "task-1129454" [ 706.784035] env[61991]: _type = "Task" [ 706.784035] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.788224] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 706.788560] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 706.788806] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Deleting the datastore file [datastore1] 527561af-ba19-447f-a4ee-618824b9ce0f {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.789481] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07139dd6-6c08-44fc-8430-891034b7054b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.797930] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129454, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.804991] env[61991]: DEBUG oslo_vmware.api [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for the task: (returnval){ [ 706.804991] env[61991]: value = "task-1129455" [ 706.804991] env[61991]: _type = "Task" [ 706.804991] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.812947] env[61991]: DEBUG oslo_vmware.api [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.845560] env[61991]: DEBUG oslo_vmware.api [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129452, 'name': PowerOffVM_Task, 'duration_secs': 0.193801} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.849449] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 706.849623] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 706.850134] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-1715711a-72c5-4e86-88e0-cecf922cb42f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.850268] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-1715711a-72c5-4e86-88e0-cecf922cb42f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.850403] env[61991]: DEBUG nova.network.neutron [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.851442] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-802fe794-d3cb-475b-be47-b21d0e5e9829 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.881393] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 706.881619] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 706.881791] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleting the datastore file [datastore1] fe78a8de-6174-47b3-9271-171c4ba76068 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.882114] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b007832-4beb-44ab-b4cc-f2faf2694640 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.889592] env[61991]: DEBUG oslo_vmware.api [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 706.889592] env[61991]: value = "task-1129457" [ 706.889592] env[61991]: _type = "Task" [ 706.889592] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.898755] env[61991]: DEBUG oslo_vmware.api [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.918805] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de4000a-a014-4af9-be24-db671a3ec404 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.926297] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808e0dff-5208-43e6-a2c0-49dd9b32fa72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.957775] env[61991]: DEBUG nova.network.neutron [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Successfully created port: 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.959840] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad15f426-c209-4cab-a42e-a199c28b3a17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.967696] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bf7104-8e59-43b0-8540-cc1b6c2d5b25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.983594] env[61991]: DEBUG nova.compute.provider_tree [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.007299] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.205352] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.293727] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129454, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060852} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.294105] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.294855] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152199c1-7f1a-4794-8b6e-cff421fb8675 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.316140] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 43cb61e7-1748-40d9-a287-1179c8219c2a/43cb61e7-1748-40d9-a287-1179c8219c2a.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.319210] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd7a9ae7-8be6-4717-b50b-fc6b8268bcbc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.338160] env[61991]: DEBUG oslo_vmware.api [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Task: {'id': task-1129455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181071} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.339298] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.339491] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 707.339666] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.339840] env[61991]: INFO nova.compute.manager [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 707.340080] env[61991]: DEBUG oslo.service.loopingcall [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.340343] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 707.340343] env[61991]: value = "task-1129458" [ 707.340343] env[61991]: _type = "Task" [ 707.340343] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.340590] env[61991]: DEBUG nova.compute.manager [-] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.340642] env[61991]: DEBUG nova.network.neutron [-] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.349836] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129458, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.396773] env[61991]: DEBUG nova.network.neutron [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.403584] env[61991]: DEBUG oslo_vmware.api [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.384824} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.403894] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.404119] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 707.404306] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.404472] env[61991]: INFO nova.compute.manager [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Took 1.09 seconds to destroy the instance on the hypervisor. [ 707.404705] env[61991]: DEBUG oslo.service.loopingcall [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.405101] env[61991]: DEBUG nova.compute.manager [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.405101] env[61991]: DEBUG nova.network.neutron [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.431493] env[61991]: DEBUG nova.network.neutron [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.486470] env[61991]: DEBUG nova.scheduler.client.report [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.552693] env[61991]: DEBUG nova.network.neutron [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Updating instance_info_cache with network_info: [{"id": "91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f", "address": "fa:16:3e:73:9c:c1", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91eaaf4e-f2", "ovs_interfaceid": "91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.702393] env[61991]: DEBUG nova.compute.manager [req-f10913c6-dd6f-4513-9fbf-a4d95a5f7be8 req-a9bd9b4d-e2e8-4621-ba68-555fc9a19119 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Received event network-vif-deleted-db700598-7ac7-4c06-9882-3206f0ae86ad {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.702393] env[61991]: INFO nova.compute.manager [req-f10913c6-dd6f-4513-9fbf-a4d95a5f7be8 req-a9bd9b4d-e2e8-4621-ba68-555fc9a19119 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Neutron deleted interface db700598-7ac7-4c06-9882-3206f0ae86ad; detaching it from the instance and deleting it from the info cache [ 707.702393] env[61991]: DEBUG nova.network.neutron [req-f10913c6-dd6f-4513-9fbf-a4d95a5f7be8 req-a9bd9b4d-e2e8-4621-ba68-555fc9a19119 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.853467] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129458, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.934196] env[61991]: DEBUG nova.network.neutron [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.991972] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.994284] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.642s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.996027] env[61991]: INFO nova.compute.claims [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.014951] env[61991]: INFO nova.scheduler.client.report [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Deleted allocations for instance a2606541-ff39-4d03-bd19-d0ee8e035c08 [ 708.020654] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.048985] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.049342] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.050527] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.050527] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.050527] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.050527] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.050527] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.050764] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.050764] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.050833] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.050953] env[61991]: DEBUG nova.virt.hardware [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.052924] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531a7eca-bbcf-484b-a677-260bf81f94f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.057353] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-1715711a-72c5-4e86-88e0-cecf922cb42f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.057627] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Instance network_info: |[{"id": "91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f", "address": "fa:16:3e:73:9c:c1", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91eaaf4e-f2", "ovs_interfaceid": "91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.060019] env[61991]: DEBUG nova.compute.manager [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Received event network-changed-91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.060019] env[61991]: DEBUG nova.compute.manager [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Refreshing instance network info cache due to event network-changed-91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.060019] env[61991]: DEBUG oslo_concurrency.lockutils [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] Acquiring lock "refresh_cache-1715711a-72c5-4e86-88e0-cecf922cb42f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.060019] env[61991]: DEBUG oslo_concurrency.lockutils [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] Acquired lock "refresh_cache-1715711a-72c5-4e86-88e0-cecf922cb42f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.060019] env[61991]: DEBUG nova.network.neutron [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Refreshing network info cache for port 91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.061758] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:9c:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3952eb02-1162-48ed-8227-9c138960d583', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.069962] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating folder: Project (e3015571d285418aae0fca50b563394e). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.071472] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f46f7e82-e441-4ad2-8a5a-b55ffb05e02d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.078232] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9520185b-d2c8-4ef8-9d2f-9b1608e5dbdb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.084455] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Created folder: Project (e3015571d285418aae0fca50b563394e) in parent group-v246753. [ 708.084455] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating folder: Instances. Parent ref: group-v246805. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.084707] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b81caf6a-2655-47d1-9394-b5d8f2d2005b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.095885] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Created folder: Instances in parent group-v246805. [ 708.096187] env[61991]: DEBUG oslo.service.loopingcall [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.096410] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.096685] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3fb4928-4209-420a-a08e-d54294e310b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.118548] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.118548] env[61991]: value = "task-1129461" [ 708.118548] env[61991]: _type = "Task" [ 708.118548] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.126999] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129461, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.157308] env[61991]: DEBUG nova.network.neutron [-] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.205940] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5281cd79-7bfb-4364-9f89-eb208bb16e19 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.214848] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcc9408-bd0b-431b-8e58-154e7711866e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.240017] env[61991]: DEBUG nova.compute.manager [req-f10913c6-dd6f-4513-9fbf-a4d95a5f7be8 req-a9bd9b4d-e2e8-4621-ba68-555fc9a19119 service nova] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Detach interface failed, port_id=db700598-7ac7-4c06-9882-3206f0ae86ad, reason: Instance 527561af-ba19-447f-a4ee-618824b9ce0f could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 708.353119] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129458, 'name': ReconfigVM_Task, 'duration_secs': 0.525477} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.353472] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 43cb61e7-1748-40d9-a287-1179c8219c2a/43cb61e7-1748-40d9-a287-1179c8219c2a.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.354228] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c68b60c-7d3b-40b9-8961-bf00c87326ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.360645] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 708.360645] env[61991]: value = "task-1129462" [ 708.360645] env[61991]: _type = "Task" [ 708.360645] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.370707] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129462, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.437671] env[61991]: INFO nova.compute.manager [-] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Took 1.03 seconds to deallocate network for instance. [ 708.491120] env[61991]: DEBUG nova.network.neutron [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Successfully updated port: 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 708.526394] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fd3309b1-dd3b-48d2-8250-793c58cd31a8 tempest-ImagesNegativeTestJSON-2053486852 tempest-ImagesNegativeTestJSON-2053486852-project-member] Lock "a2606541-ff39-4d03-bd19-d0ee8e035c08" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.613s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.628281] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129461, 'name': CreateVM_Task, 'duration_secs': 0.331456} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.628447] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 708.629149] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.629341] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.629658] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 708.629905] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfdc0dc3-64c1-49c2-8472-c6dbd0899177 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.634336] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 708.634336] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ef1ecb-cb9b-520b-4f84-abd266904c2c" [ 708.634336] env[61991]: _type = "Task" [ 708.634336] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.642231] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ef1ecb-cb9b-520b-4f84-abd266904c2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.662022] env[61991]: INFO nova.compute.manager [-] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Took 1.32 seconds to deallocate network for instance. [ 708.842546] env[61991]: DEBUG nova.network.neutron [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Updated VIF entry in instance network info cache for port 91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 708.842813] env[61991]: DEBUG nova.network.neutron [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Updating instance_info_cache with network_info: [{"id": "91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f", "address": "fa:16:3e:73:9c:c1", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91eaaf4e-f2", "ovs_interfaceid": "91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.870739] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129462, 'name': Rename_Task, 'duration_secs': 0.14113} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.871059] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 708.871308] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b1cf6fb-28c7-4e10-b1d6-8be1983eeab2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.877157] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 708.877157] env[61991]: value = "task-1129463" [ 708.877157] env[61991]: _type = "Task" [ 708.877157] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.884508] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.943559] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.993262] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.993425] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.993616] env[61991]: DEBUG nova.network.neutron [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.151235] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ef1ecb-cb9b-520b-4f84-abd266904c2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010122} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.151535] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.151757] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.152777] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.152777] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.152777] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.152777] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b56d4303-525a-4bf7-81b2-b94f6ca4a1fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.165353] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.165628] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 709.168475] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b5adac1-4a49-4eed-a398-26bf7b64c0a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.171412] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.176691] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 709.176691] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b64288-7b43-7f93-78f4-f4ad0baa0bec" [ 709.176691] env[61991]: _type = "Task" [ 709.176691] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.187608] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b64288-7b43-7f93-78f4-f4ad0baa0bec, 'name': SearchDatastore_Task, 'duration_secs': 0.007972} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.188394] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bb65d99-b8ae-4a27-a6e3-5a864a563550 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.193298] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 709.193298] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52730a72-4ab3-4364-b97c-596d001f6878" [ 709.193298] env[61991]: _type = "Task" [ 709.193298] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.204099] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52730a72-4ab3-4364-b97c-596d001f6878, 'name': SearchDatastore_Task, 'duration_secs': 0.008182} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.207906] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.208215] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 1715711a-72c5-4e86-88e0-cecf922cb42f/1715711a-72c5-4e86-88e0-cecf922cb42f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.211029] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03c4b7b5-a90c-41cf-85da-06035751946f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.217610] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 709.217610] env[61991]: value = "task-1129464" [ 709.217610] env[61991]: _type = "Task" [ 709.217610] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.225602] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.346452] env[61991]: DEBUG oslo_concurrency.lockutils [req-48037312-3461-4165-a36e-14b0c3e86956 req-88a8f81f-8478-4463-b12d-f03c421bccc1 service nova] Releasing lock "refresh_cache-1715711a-72c5-4e86-88e0-cecf922cb42f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.393046] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129463, 'name': PowerOnVM_Task} progress is 19%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.408539] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c34f00-4530-4019-860b-a047c466a5de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.418537] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06472af0-721e-4b37-9dda-0a1093a6ebf0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.452775] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a4f727-bb94-4937-baa9-4464cd4119d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.461862] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54537883-90c0-4945-a209-657d839c689e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.476664] env[61991]: DEBUG nova.compute.provider_tree [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.530068] env[61991]: DEBUG nova.network.neutron [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.710327] env[61991]: DEBUG nova.network.neutron [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updating instance_info_cache with network_info: [{"id": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "address": "fa:16:3e:8e:7c:67", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fdca45f-c9", "ovs_interfaceid": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.730073] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498467} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.730423] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 1715711a-72c5-4e86-88e0-cecf922cb42f/1715711a-72c5-4e86-88e0-cecf922cb42f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 709.730668] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.730958] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8aba802c-7a81-4a64-8672-9068d3d033e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.737897] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 709.737897] env[61991]: value = "task-1129465" [ 709.737897] env[61991]: _type = "Task" [ 709.737897] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.746281] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.800644] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.800896] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.890296] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129463, 'name': PowerOnVM_Task} progress is 64%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.981899] env[61991]: DEBUG nova.scheduler.client.report [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.163492] env[61991]: DEBUG nova.compute.manager [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Received event network-vif-plugged-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.163817] env[61991]: DEBUG oslo_concurrency.lockutils [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] Acquiring lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.163979] env[61991]: DEBUG oslo_concurrency.lockutils [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.164176] env[61991]: DEBUG oslo_concurrency.lockutils [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.164345] env[61991]: DEBUG nova.compute.manager [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] No waiting events found dispatching network-vif-plugged-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 710.164561] env[61991]: WARNING nova.compute.manager [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Received unexpected event network-vif-plugged-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 for instance with vm_state building and task_state spawning. [ 710.164722] env[61991]: DEBUG nova.compute.manager [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Received event network-changed-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.164877] env[61991]: DEBUG nova.compute.manager [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Refreshing instance network info cache due to event network-changed-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.165051] env[61991]: DEBUG oslo_concurrency.lockutils [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] Acquiring lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.212698] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.213105] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Instance network_info: |[{"id": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "address": "fa:16:3e:8e:7c:67", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fdca45f-c9", "ovs_interfaceid": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 710.213407] env[61991]: DEBUG oslo_concurrency.lockutils [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] Acquired lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.213581] env[61991]: DEBUG nova.network.neutron [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Refreshing network info cache for port 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.219033] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:7c:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 710.225071] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating folder: Project (adbd4ea9b0744f28a0d4a46b4a04d683). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 710.226253] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-132de55e-60f9-4fc8-937b-daaf0adc38c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.237775] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created folder: Project (adbd4ea9b0744f28a0d4a46b4a04d683) in parent group-v246753. [ 710.237994] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating folder: Instances. Parent ref: group-v246808. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 710.238578] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b8470eb-0a8c-49b7-ac3f-88fb7dbc7c93 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.252118] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061894} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.252700] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.253196] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dcfe11-dc77-493a-ad4a-64899d9d7323 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.256695] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created folder: Instances in parent group-v246808. [ 710.256913] env[61991]: DEBUG oslo.service.loopingcall [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.257412] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 710.257609] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ebaa538-1935-499b-99dc-c5f95c2fd294 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.290389] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 1715711a-72c5-4e86-88e0-cecf922cb42f/1715711a-72c5-4e86-88e0-cecf922cb42f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.291265] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-339258e3-aacd-46b5-ba40-f48662b63219 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.311163] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 710.311163] env[61991]: value = "task-1129468" [ 710.311163] env[61991]: _type = "Task" [ 710.311163] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.315848] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 710.316049] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 710.316164] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 710.319211] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 710.319211] env[61991]: value = "task-1129469" [ 710.319211] env[61991]: _type = "Task" [ 710.319211] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.325959] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129468, 'name': CreateVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.332252] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129469, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.387809] env[61991]: DEBUG oslo_vmware.api [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129463, 'name': PowerOnVM_Task, 'duration_secs': 1.189486} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.388139] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 710.388370] env[61991]: INFO nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Took 11.30 seconds to spawn the instance on the hypervisor. [ 710.388602] env[61991]: DEBUG nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 710.389392] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52277524-ab5a-446c-bad4-98fde17f57a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.489588] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.490203] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.493189] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.522s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.493452] env[61991]: DEBUG nova.objects.instance [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lazy-loading 'resources' on Instance uuid 9d88002b-d5f0-4669-9a34-bfbf8f9169d3 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 710.823602] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 710.823769] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 710.823923] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 710.824095] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 710.826184] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129468, 'name': CreateVM_Task, 'duration_secs': 0.401076} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.833148] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 710.834354] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.834462] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.834871] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 710.835621] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ad288e8-9eb4-46e4-ac91-52933cc8022c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.840151] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129469, 'name': ReconfigVM_Task, 'duration_secs': 0.311763} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.840721] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 1715711a-72c5-4e86-88e0-cecf922cb42f/1715711a-72c5-4e86-88e0-cecf922cb42f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.841331] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9a78b47-2444-4f8c-872d-48910ff04643 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.843899] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 710.843899] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5263e057-327d-4f22-afc3-8016997686d8" [ 710.843899] env[61991]: _type = "Task" [ 710.843899] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.848059] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 710.848059] env[61991]: value = "task-1129470" [ 710.848059] env[61991]: _type = "Task" [ 710.848059] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.853948] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5263e057-327d-4f22-afc3-8016997686d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.855879] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.856016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.856163] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 710.856314] env[61991]: DEBUG nova.objects.instance [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lazy-loading 'info_cache' on Instance uuid 9d88002b-d5f0-4669-9a34-bfbf8f9169d3 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 710.862118] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129470, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.906203] env[61991]: INFO nova.compute.manager [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Took 35.99 seconds to build instance. [ 710.983261] env[61991]: DEBUG nova.network.neutron [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updated VIF entry in instance network info cache for port 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 710.983689] env[61991]: DEBUG nova.network.neutron [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updating instance_info_cache with network_info: [{"id": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "address": "fa:16:3e:8e:7c:67", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fdca45f-c9", "ovs_interfaceid": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.996736] env[61991]: DEBUG nova.compute.utils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.000695] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.000865] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.053350] env[61991]: DEBUG nova.policy [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69e4ad858d634a5789edf63ccd01faac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8446aad15ca847c19507c1b44b64b304', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.362799] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5263e057-327d-4f22-afc3-8016997686d8, 'name': SearchDatastore_Task, 'duration_secs': 0.030382} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.364922] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129470, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.365490] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.365738] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 711.365962] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.366121] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.366297] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 711.366535] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2607eeb4-0836-4779-b44f-cb01e3176f1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.375954] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d570209-4b89-4a66-a317-1f6f4df206c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.380209] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 711.380388] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 711.381818] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad20ff10-4ab1-4833-9094-8046a8189c93 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.387229] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c08718a-a2c2-4911-b6a0-bda6762fd948 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.391637] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 711.391637] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520ebbfa-3cf0-5cc7-65ea-73eaf25d9366" [ 711.391637] env[61991]: _type = "Task" [ 711.391637] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.421672] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b195e67f-6d5f-45bd-b806-d6f62b4555ad tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.202s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.426058] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dec5ce-e016-4d5a-845e-edc7776a03ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.430328] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520ebbfa-3cf0-5cc7-65ea-73eaf25d9366, 'name': SearchDatastore_Task, 'duration_secs': 0.018964} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.431978] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8810abe4-e622-4b42-9841-a8d1b1edeabc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.437279] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469b3c2b-584c-4b0e-a8ab-b5666b29c1e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.444316] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 711.444316] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fc4f10-11a4-c972-4b28-4defc1722700" [ 711.444316] env[61991]: _type = "Task" [ 711.444316] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.456174] env[61991]: DEBUG nova.compute.provider_tree [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.462693] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fc4f10-11a4-c972-4b28-4defc1722700, 'name': SearchDatastore_Task, 'duration_secs': 0.016249} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.463540] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Successfully created port: 4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.465055] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.465317] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] cf99c9cc-24c3-4acc-8120-49c4b12a3553/cf99c9cc-24c3-4acc-8120-49c4b12a3553.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 711.465569] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b37925e1-92d3-41d0-bbf3-023ea01fade9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.471768] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 711.471768] env[61991]: value = "task-1129471" [ 711.471768] env[61991]: _type = "Task" [ 711.471768] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.479966] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.486923] env[61991]: DEBUG oslo_concurrency.lockutils [req-c712450f-5bb1-4c05-b9e4-45ebb3484533 req-3e2a752e-6a25-4c4e-b58a-6e1525177177 service nova] Releasing lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.504287] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.782670] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Successfully created port: 15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.860167] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129470, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.891804] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.928201] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.960837] env[61991]: DEBUG nova.scheduler.client.report [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.986945] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129471, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.239973] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "43cb61e7-1748-40d9-a287-1179c8219c2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.240097] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.240365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "43cb61e7-1748-40d9-a287-1179c8219c2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.240607] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.241032] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.243754] env[61991]: INFO nova.compute.manager [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Terminating instance [ 712.245934] env[61991]: DEBUG nova.compute.manager [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.246206] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.247139] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ae422d-10a9-4745-a4a6-813a12170631 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.256010] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 712.256266] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5713a3bb-1e91-4c96-9db4-d4422c3a6402 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.266901] env[61991]: DEBUG oslo_vmware.api [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 712.266901] env[61991]: value = "task-1129472" [ 712.266901] env[61991]: _type = "Task" [ 712.266901] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.274179] env[61991]: DEBUG oslo_vmware.api [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.359591] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129470, 'name': Rename_Task, 'duration_secs': 1.395738} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.359907] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 712.360142] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eeae708a-cf87-48d8-a80c-635804a2cbb0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.369023] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 712.369023] env[61991]: value = "task-1129473" [ 712.369023] env[61991]: _type = "Task" [ 712.369023] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.376407] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.453687] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.469375] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.471718] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.261s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.473767] env[61991]: INFO nova.compute.claims [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.485841] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129471, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760484} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.488938] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] cf99c9cc-24c3-4acc-8120-49c4b12a3553/cf99c9cc-24c3-4acc-8120-49c4b12a3553.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 712.488938] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.488938] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d1249af-4b0e-413e-aecb-5576672e8f07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.489853] env[61991]: INFO nova.scheduler.client.report [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Deleted allocations for instance 9d88002b-d5f0-4669-9a34-bfbf8f9169d3 [ 712.497662] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 712.497662] env[61991]: value = "task-1129474" [ 712.497662] env[61991]: _type = "Task" [ 712.497662] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.506198] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.513407] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.532032] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.550575] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.550821] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.550984] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.551175] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.551316] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.551489] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.551693] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.551904] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.552685] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.552685] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.552685] env[61991]: DEBUG nova.virt.hardware [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.553694] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd440b98-1d0f-45b8-bbe7-d7c692ed6d5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.563976] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f2763e-3026-4d9b-b572-5974b3a934ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.777337] env[61991]: DEBUG oslo_vmware.api [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129472, 'name': PowerOffVM_Task, 'duration_secs': 0.191534} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.777575] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 712.777718] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 712.777972] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef627ab7-50cb-4cb4-b456-efa39dbf409d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.846239] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 712.846239] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 712.846239] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Deleting the datastore file [datastore2] 43cb61e7-1748-40d9-a287-1179c8219c2a {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 712.846239] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65cee15d-e607-40e2-a660-174c77628082 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.853727] env[61991]: DEBUG oslo_vmware.api [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for the task: (returnval){ [ 712.853727] env[61991]: value = "task-1129476" [ 712.853727] env[61991]: _type = "Task" [ 712.853727] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.866781] env[61991]: DEBUG oslo_vmware.api [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.880030] env[61991]: DEBUG oslo_vmware.api [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129473, 'name': PowerOnVM_Task, 'duration_secs': 0.505379} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.880030] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 712.880030] env[61991]: INFO nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Took 7.35 seconds to spawn the instance on the hypervisor. [ 712.880030] env[61991]: DEBUG nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 712.881352] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a916f2-f5be-4504-a157-04a4e3e408dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.002271] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ed91a0d-1469-47bc-81f0-0f75bb5f5b89 tempest-ServerMetadataNegativeTestJSON-1553920805 tempest-ServerMetadataNegativeTestJSON-1553920805-project-member] Lock "9d88002b-d5f0-4669-9a34-bfbf8f9169d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.968s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.012651] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065425} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.012923] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 713.013738] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f21814d-c276-4713-bd0d-f9b2941d54a6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.039789] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] cf99c9cc-24c3-4acc-8120-49c4b12a3553/cf99c9cc-24c3-4acc-8120-49c4b12a3553.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 713.040875] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-9d88002b-d5f0-4669-9a34-bfbf8f9169d3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.041126] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 713.041348] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d37b38e0-ac37-4757-8954-6c2f4d43e4f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.055612] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.056204] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.056820] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.057035] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.057263] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.057426] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.057584] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 713.057757] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 713.061553] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 713.061553] env[61991]: value = "task-1129477" [ 713.061553] env[61991]: _type = "Task" [ 713.061553] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.073330] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129477, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.364545] env[61991]: DEBUG oslo_vmware.api [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Task: {'id': task-1129476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16177} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.365072] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 713.365375] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 713.365655] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.365962] env[61991]: INFO nova.compute.manager [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 713.366317] env[61991]: DEBUG oslo.service.loopingcall [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.366640] env[61991]: DEBUG nova.compute.manager [-] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.366833] env[61991]: DEBUG nova.network.neutron [-] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.405081] env[61991]: INFO nova.compute.manager [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Took 33.99 seconds to build instance. [ 713.561264] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.576125] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129477, 'name': ReconfigVM_Task, 'duration_secs': 0.267135} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.576125] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Reconfigured VM instance instance-0000002f to attach disk [datastore2] cf99c9cc-24c3-4acc-8120-49c4b12a3553/cf99c9cc-24c3-4acc-8120-49c4b12a3553.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.576436] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09072597-167a-4f9e-b4b7-4ae873d6fe63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.587016] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 713.587016] env[61991]: value = "task-1129478" [ 713.587016] env[61991]: _type = "Task" [ 713.587016] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.596800] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129478, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.836179] env[61991]: DEBUG nova.compute.manager [req-a5f9ece6-10b9-4966-a113-e774c1c3b772 req-7b155bcd-7c5b-4254-b5da-163af2df10c6 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received event network-vif-plugged-4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.836395] env[61991]: DEBUG oslo_concurrency.lockutils [req-a5f9ece6-10b9-4966-a113-e774c1c3b772 req-7b155bcd-7c5b-4254-b5da-163af2df10c6 service nova] Acquiring lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.836590] env[61991]: DEBUG oslo_concurrency.lockutils [req-a5f9ece6-10b9-4966-a113-e774c1c3b772 req-7b155bcd-7c5b-4254-b5da-163af2df10c6 service nova] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.836748] env[61991]: DEBUG oslo_concurrency.lockutils [req-a5f9ece6-10b9-4966-a113-e774c1c3b772 req-7b155bcd-7c5b-4254-b5da-163af2df10c6 service nova] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.836919] env[61991]: DEBUG nova.compute.manager [req-a5f9ece6-10b9-4966-a113-e774c1c3b772 req-7b155bcd-7c5b-4254-b5da-163af2df10c6 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] No waiting events found dispatching network-vif-plugged-4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 713.837087] env[61991]: WARNING nova.compute.manager [req-a5f9ece6-10b9-4966-a113-e774c1c3b772 req-7b155bcd-7c5b-4254-b5da-163af2df10c6 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received unexpected event network-vif-plugged-4c68db77-3ed0-4069-ab22-78bb0f625e1c for instance with vm_state building and task_state spawning. [ 713.857397] env[61991]: DEBUG nova.compute.manager [req-bcb43be7-65cd-4db6-999d-43f11466e589 req-1b45a88a-87d3-4b48-aea0-c2cd5542ed7c service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Received event network-vif-deleted-72dca87d-ec73-4972-8883-0f9dd0284313 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.857397] env[61991]: INFO nova.compute.manager [req-bcb43be7-65cd-4db6-999d-43f11466e589 req-1b45a88a-87d3-4b48-aea0-c2cd5542ed7c service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Neutron deleted interface 72dca87d-ec73-4972-8883-0f9dd0284313; detaching it from the instance and deleting it from the info cache [ 713.857532] env[61991]: DEBUG nova.network.neutron [req-bcb43be7-65cd-4db6-999d-43f11466e589 req-1b45a88a-87d3-4b48-aea0-c2cd5542ed7c service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.893848] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Successfully updated port: 4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 713.897594] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2d1918-8863-4a5b-a81e-aa8e78f0dbbb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.905706] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67ae45b-7d9a-42c3-a1aa-8eeb33f62c77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.909774] env[61991]: DEBUG oslo_concurrency.lockutils [None req-361250f0-1ea3-40c9-8d64-0b208ea27ea6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.860s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.951669] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3755b9-e56a-4ed6-ac29-91f0660ab9e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.960267] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aaa55f-1556-407a-88d4-046959a29d77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.975369] env[61991]: DEBUG nova.compute.provider_tree [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.095645] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129478, 'name': Rename_Task, 'duration_secs': 0.140245} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.095920] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 714.096188] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4d34133-dc03-47c6-9da1-1d33b9f25f71 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.101972] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 714.101972] env[61991]: value = "task-1129479" [ 714.101972] env[61991]: _type = "Task" [ 714.101972] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.109571] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.168232] env[61991]: DEBUG nova.network.neutron [-] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.361131] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7b2e1481-5013-4020-9984-838f3502cbc4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.371627] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d110b6-924e-44d7-b932-e43d80efac0d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.397822] env[61991]: DEBUG nova.compute.manager [req-bcb43be7-65cd-4db6-999d-43f11466e589 req-1b45a88a-87d3-4b48-aea0-c2cd5542ed7c service nova] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Detach interface failed, port_id=72dca87d-ec73-4972-8883-0f9dd0284313, reason: Instance 43cb61e7-1748-40d9-a287-1179c8219c2a could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 714.412695] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.481317] env[61991]: DEBUG nova.scheduler.client.report [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.613931] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129479, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.672349] env[61991]: INFO nova.compute.manager [-] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Took 1.31 seconds to deallocate network for instance. [ 714.940616] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.988477] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.988927] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.992220] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.728s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.993828] env[61991]: INFO nova.compute.claims [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.112920] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129479, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.180807] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.455817] env[61991]: DEBUG nova.compute.manager [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.456854] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc0dd73-3f7c-4973-a7d2-6c80680bd36e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.494764] env[61991]: DEBUG nova.compute.utils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.496772] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.496938] env[61991]: DEBUG nova.network.neutron [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.554969] env[61991]: DEBUG nova.policy [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8946c3371164af6b2943fddd282f977', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7b25d4745e1419b963e387c8461dd7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.614056] env[61991]: DEBUG oslo_vmware.api [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129479, 'name': PowerOnVM_Task, 'duration_secs': 1.018606} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.614354] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 715.614580] env[61991]: INFO nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Took 7.59 seconds to spawn the instance on the hypervisor. [ 715.614725] env[61991]: DEBUG nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.615647] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3b7462-abf5-499b-8694-809174ccc104 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.899061] env[61991]: DEBUG nova.network.neutron [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Successfully created port: 4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.907644] env[61991]: DEBUG nova.compute.manager [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received event network-changed-4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.907867] env[61991]: DEBUG nova.compute.manager [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Refreshing instance network info cache due to event network-changed-4c68db77-3ed0-4069-ab22-78bb0f625e1c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 715.908093] env[61991]: DEBUG oslo_concurrency.lockutils [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] Acquiring lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.908235] env[61991]: DEBUG oslo_concurrency.lockutils [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] Acquired lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.908388] env[61991]: DEBUG nova.network.neutron [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Refreshing network info cache for port 4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 715.967830] env[61991]: INFO nova.compute.manager [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] instance snapshotting [ 715.970736] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e16b633-573f-4651-be3e-f90df1476c4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.993849] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fa4c05-d832-4731-9bb3-df90d56d5ce9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.001240] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.141132] env[61991]: INFO nova.compute.manager [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Took 36.35 seconds to build instance. [ 716.159955] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Successfully updated port: 15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 716.414080] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455b36c2-f702-4cf4-a310-d4b64502b72b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.428591] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca2de01-b717-4035-be0b-3f7c57f5f38f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.460581] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e060d8a-1d9b-44ed-89fc-91fc0b8c4e8f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.468084] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ae8a07-490a-4405-ad55-05ba2feea646 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.482710] env[61991]: DEBUG nova.compute.provider_tree [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.514028] env[61991]: DEBUG nova.network.neutron [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.514028] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 716.514028] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-65dd5290-3e86-4e92-8980-063c0ff44cd2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.521730] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 716.521730] env[61991]: value = "task-1129480" [ 716.521730] env[61991]: _type = "Task" [ 716.521730] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.530772] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129480, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.643835] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5a601bea-196d-4087-ba0f-48d8cafc287a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.712s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.669204] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.892895] env[61991]: DEBUG nova.network.neutron [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.985187] env[61991]: DEBUG nova.scheduler.client.report [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.015600] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.032025] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129480, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.046619] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.046861] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.047024] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.047211] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.047353] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.047498] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.047840] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.048013] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.048269] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.048437] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.048605] env[61991]: DEBUG nova.virt.hardware [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.049424] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60655435-0643-4ef1-82ee-1760e2770e9b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.057131] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9c821b-d581-44ef-8e31-6ac55f033a29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.147154] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.400013] env[61991]: DEBUG oslo_concurrency.lockutils [req-1ca9d78b-2f69-4d98-9afa-630897c14139 req-947436be-8f1f-4885-b902-6c88a10aa7cb service nova] Releasing lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.400013] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquired lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.400013] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.454823] env[61991]: DEBUG nova.compute.manager [req-3d63e5a4-9e5d-42b0-a451-66747d167f2f req-0c04034f-f8fc-4167-b0cb-364da6c8c6cb service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-vif-plugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.456257] env[61991]: DEBUG oslo_concurrency.lockutils [req-3d63e5a4-9e5d-42b0-a451-66747d167f2f req-0c04034f-f8fc-4167-b0cb-364da6c8c6cb service nova] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.456606] env[61991]: DEBUG oslo_concurrency.lockutils [req-3d63e5a4-9e5d-42b0-a451-66747d167f2f req-0c04034f-f8fc-4167-b0cb-364da6c8c6cb service nova] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.457185] env[61991]: DEBUG oslo_concurrency.lockutils [req-3d63e5a4-9e5d-42b0-a451-66747d167f2f req-0c04034f-f8fc-4167-b0cb-364da6c8c6cb service nova] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.457539] env[61991]: DEBUG nova.compute.manager [req-3d63e5a4-9e5d-42b0-a451-66747d167f2f req-0c04034f-f8fc-4167-b0cb-364da6c8c6cb service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] No waiting events found dispatching network-vif-plugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 717.457806] env[61991]: WARNING nova.compute.manager [req-3d63e5a4-9e5d-42b0-a451-66747d167f2f req-0c04034f-f8fc-4167-b0cb-364da6c8c6cb service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received unexpected event network-vif-plugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 for instance with vm_state building and task_state spawning. [ 717.493020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.493020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.580s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.494171] env[61991]: INFO nova.compute.claims [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.532223] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129480, 'name': CreateSnapshot_Task, 'duration_secs': 0.549003} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.532506] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 717.533257] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6db261-e3d7-4f98-a3e1-1a0b9a3a774c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.563618] env[61991]: DEBUG nova.network.neutron [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Successfully updated port: 4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 717.669772] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.942648] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.997950] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "5774d63b-f527-429f-bbd1-07d46fb09b2b" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.998373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "5774d63b-f527-429f-bbd1-07d46fb09b2b" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.028142] env[61991]: DEBUG nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received event network-vif-plugged-15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.028371] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Acquiring lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.028596] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.029016] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.029016] env[61991]: DEBUG nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] No waiting events found dispatching network-vif-plugged-15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 718.029169] env[61991]: WARNING nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received unexpected event network-vif-plugged-15c923b4-4e91-4367-9346-cb8b6e51c275 for instance with vm_state building and task_state spawning. [ 718.029300] env[61991]: DEBUG nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received event network-changed-15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.029436] env[61991]: DEBUG nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Refreshing instance network info cache due to event network-changed-15c923b4-4e91-4367-9346-cb8b6e51c275. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.029643] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Acquiring lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.052277] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 718.052845] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-867300c9-6824-493f-aec2-f6d5c8453907 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.060905] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 718.060905] env[61991]: value = "task-1129481" [ 718.060905] env[61991]: _type = "Task" [ 718.060905] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.066712] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.066787] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.066861] env[61991]: DEBUG nova.network.neutron [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.072852] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129481, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.341274] env[61991]: DEBUG nova.network.neutron [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Updating instance_info_cache with network_info: [{"id": "4c68db77-3ed0-4069-ab22-78bb0f625e1c", "address": "fa:16:3e:6e:11:9a", "network": {"id": "41637242-c208-4ddd-b98d-2c16c0703b4a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1012001284", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8446aad15ca847c19507c1b44b64b304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c68db77-3e", "ovs_interfaceid": "4c68db77-3ed0-4069-ab22-78bb0f625e1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15c923b4-4e91-4367-9346-cb8b6e51c275", "address": "fa:16:3e:58:21:c7", "network": {"id": "29794684-f9f9-4099-84fc-bd74f9025795", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1022442373", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.223", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8446aad15ca847c19507c1b44b64b304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5756d009-13ad-4e13-a991-3b5e71830aa5", "external-id": "nsx-vlan-transportzone-608", "segmentation_id": 608, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15c923b4-4e", "ovs_interfaceid": "15c923b4-4e91-4367-9346-cb8b6e51c275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.503661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "5774d63b-f527-429f-bbd1-07d46fb09b2b" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.506622] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.572510] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129481, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.603135] env[61991]: DEBUG nova.network.neutron [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.747903] env[61991]: DEBUG nova.network.neutron [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.809849] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8c7e92-c1c1-4c81-b183-41c16fa2be1a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.817627] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83eb2e5f-6a16-45ff-b352-4c4a76150d6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.847605] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Releasing lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.847959] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Instance network_info: |[{"id": "4c68db77-3ed0-4069-ab22-78bb0f625e1c", "address": "fa:16:3e:6e:11:9a", "network": {"id": "41637242-c208-4ddd-b98d-2c16c0703b4a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1012001284", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8446aad15ca847c19507c1b44b64b304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c68db77-3e", "ovs_interfaceid": "4c68db77-3ed0-4069-ab22-78bb0f625e1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15c923b4-4e91-4367-9346-cb8b6e51c275", "address": "fa:16:3e:58:21:c7", "network": {"id": "29794684-f9f9-4099-84fc-bd74f9025795", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1022442373", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.223", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8446aad15ca847c19507c1b44b64b304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5756d009-13ad-4e13-a991-3b5e71830aa5", "external-id": "nsx-vlan-transportzone-608", "segmentation_id": 608, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15c923b4-4e", "ovs_interfaceid": "15c923b4-4e91-4367-9346-cb8b6e51c275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 718.848521] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Acquired lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.848699] env[61991]: DEBUG nova.network.neutron [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Refreshing network info cache for port 15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.849830] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:11:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55bd18a7-39a8-4d07-9088-9b944f9ff710', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c68db77-3ed0-4069-ab22-78bb0f625e1c', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:21:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5756d009-13ad-4e13-a991-3b5e71830aa5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15c923b4-4e91-4367-9346-cb8b6e51c275', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.858766] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Creating folder: Project (8446aad15ca847c19507c1b44b64b304). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.859630] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a504d08-ac67-4f5b-a9b3-2027209ba752 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.864809] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ea4fe35-f6f0-4ae1-82e3-7b7961186a23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.871640] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a8d744-38c1-461d-9ea4-9423d910986d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.877746] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Created folder: Project (8446aad15ca847c19507c1b44b64b304) in parent group-v246753. [ 718.877906] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Creating folder: Instances. Parent ref: group-v246813. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.878405] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4eb0cb4-7de6-4ee0-932c-40e2b96c8e66 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.888085] env[61991]: DEBUG nova.compute.provider_tree [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.889773] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Created folder: Instances in parent group-v246813. [ 718.889985] env[61991]: DEBUG oslo.service.loopingcall [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.890362] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 718.890557] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4bca7f7-679d-463e-be61-39c50244a4be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.912785] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.912785] env[61991]: value = "task-1129484" [ 718.912785] env[61991]: _type = "Task" [ 718.912785] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.920757] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129484, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.013462] env[61991]: DEBUG nova.compute.utils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.014955] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.015147] env[61991]: DEBUG nova.network.neutron [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.074994] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129481, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.076669] env[61991]: DEBUG nova.policy [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11957ba76327421ca67414fcbea11368', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be45d020174c488b9a55e45bed5ac324', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.105678] env[61991]: DEBUG nova.network.neutron [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Updated VIF entry in instance network info cache for port 15c923b4-4e91-4367-9346-cb8b6e51c275. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 719.106177] env[61991]: DEBUG nova.network.neutron [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Updating instance_info_cache with network_info: [{"id": "4c68db77-3ed0-4069-ab22-78bb0f625e1c", "address": "fa:16:3e:6e:11:9a", "network": {"id": "41637242-c208-4ddd-b98d-2c16c0703b4a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1012001284", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8446aad15ca847c19507c1b44b64b304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55bd18a7-39a8-4d07-9088-9b944f9ff710", "external-id": "nsx-vlan-transportzone-686", "segmentation_id": 686, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c68db77-3e", "ovs_interfaceid": "4c68db77-3ed0-4069-ab22-78bb0f625e1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15c923b4-4e91-4367-9346-cb8b6e51c275", "address": "fa:16:3e:58:21:c7", "network": {"id": "29794684-f9f9-4099-84fc-bd74f9025795", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1022442373", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.223", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "8446aad15ca847c19507c1b44b64b304", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5756d009-13ad-4e13-a991-3b5e71830aa5", "external-id": "nsx-vlan-transportzone-608", "segmentation_id": 608, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15c923b4-4e", "ovs_interfaceid": "15c923b4-4e91-4367-9346-cb8b6e51c275", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.252908] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.253423] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance network_info: |[{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 719.253865] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:89:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b255035-ebfa-4e8a-b9e5-a51737f080b9', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 719.261416] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating folder: Project (d7b25d4745e1419b963e387c8461dd7d). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 719.261751] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee8da997-c1cf-48dd-8bac-62aade7ccfb5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.271637] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Created folder: Project (d7b25d4745e1419b963e387c8461dd7d) in parent group-v246753. [ 719.271856] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating folder: Instances. Parent ref: group-v246816. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 719.272081] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09079a7a-7245-4a67-8e21-6f01ec440652 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.280299] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Created folder: Instances in parent group-v246816. [ 719.280529] env[61991]: DEBUG oslo.service.loopingcall [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.280716] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 719.280909] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9affa3ae-ebeb-4317-a361-8f7067dbdf32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.300576] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 719.300576] env[61991]: value = "task-1129487" [ 719.300576] env[61991]: _type = "Task" [ 719.300576] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.308513] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129487, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.391376] env[61991]: DEBUG nova.scheduler.client.report [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.430981] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129484, 'name': CreateVM_Task, 'duration_secs': 0.341239} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.430981] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.432360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.432360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.432360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 719.432697] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dee44746-8459-4de8-86d7-bdd8ac77b65d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.437474] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 719.437474] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527c9478-0388-7ddb-0f20-96620943357f" [ 719.437474] env[61991]: _type = "Task" [ 719.437474] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.438501] env[61991]: DEBUG nova.network.neutron [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Successfully created port: 227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.452033] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527c9478-0388-7ddb-0f20-96620943357f, 'name': SearchDatastore_Task, 'duration_secs': 0.00912} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.452033] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.452033] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.452327] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.452521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.452741] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.453041] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-687884a1-f11f-4517-96b2-3538ca00f5fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.462226] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.462505] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 719.463463] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d6c1a49-875c-4f33-8e03-772a035c60cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.469547] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 719.469547] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287b058-ee63-9d0b-e4a8-f6cee44872a6" [ 719.469547] env[61991]: _type = "Task" [ 719.469547] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.477973] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287b058-ee63-9d0b-e4a8-f6cee44872a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.523334] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.573908] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129481, 'name': CloneVM_Task, 'duration_secs': 1.503875} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.574149] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Created linked-clone VM from snapshot [ 719.575066] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb87089-2bdb-4d6d-b941-7cbaaf1872bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.584096] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Uploading image e4b0de21-26e3-4d65-834b-7a893af37f7a {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 719.608500] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 719.608500] env[61991]: value = "vm-246812" [ 719.608500] env[61991]: _type = "VirtualMachine" [ 719.608500] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 719.609018] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Releasing lock "refresh_cache-0631634f-5c58-4ab0-b0dd-71821caf2fc6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.609267] env[61991]: DEBUG nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Received event network-changed-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.609429] env[61991]: DEBUG nova.compute.manager [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Refreshing instance network info cache due to event network-changed-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.609626] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Acquiring lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.609764] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Acquired lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.609920] env[61991]: DEBUG nova.network.neutron [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Refreshing network info cache for port 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 719.611538] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1ab796c6-d2d7-4c76-9107-085bcc2fc9b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.615641] env[61991]: DEBUG nova.compute.manager [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.615774] env[61991]: DEBUG nova.compute.manager [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing instance network info cache due to event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.615974] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.616128] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.616288] env[61991]: DEBUG nova.network.neutron [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 719.623279] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lease: (returnval){ [ 719.623279] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cc074f-4742-0d9f-9642-c4cc781292b8" [ 719.623279] env[61991]: _type = "HttpNfcLease" [ 719.623279] env[61991]: } obtained for exporting VM: (result){ [ 719.623279] env[61991]: value = "vm-246812" [ 719.623279] env[61991]: _type = "VirtualMachine" [ 719.623279] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 719.623865] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the lease: (returnval){ [ 719.623865] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cc074f-4742-0d9f-9642-c4cc781292b8" [ 719.623865] env[61991]: _type = "HttpNfcLease" [ 719.623865] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 719.634275] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 719.634275] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cc074f-4742-0d9f-9642-c4cc781292b8" [ 719.634275] env[61991]: _type = "HttpNfcLease" [ 719.634275] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 719.810517] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129487, 'name': CreateVM_Task, 'duration_secs': 0.291952} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.810686] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.811423] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.811589] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.811882] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 719.812159] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf8496ef-f6ab-4b4e-8de4-e3697fdea3fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.816339] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 719.816339] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521f4dc0-07d2-21de-1378-14bc0c644a13" [ 719.816339] env[61991]: _type = "Task" [ 719.816339] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.823578] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521f4dc0-07d2-21de-1378-14bc0c644a13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.898584] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.899123] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.901752] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.634s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.903237] env[61991]: INFO nova.compute.claims [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.979635] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287b058-ee63-9d0b-e4a8-f6cee44872a6, 'name': SearchDatastore_Task, 'duration_secs': 0.00879} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.980415] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-279c4e28-377d-434f-885d-c87a21516009 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.985267] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 719.985267] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bcad1d-93f3-74e8-c683-447d36eda9e3" [ 719.985267] env[61991]: _type = "Task" [ 719.985267] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.992088] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bcad1d-93f3-74e8-c683-447d36eda9e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.133496] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 720.133496] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cc074f-4742-0d9f-9642-c4cc781292b8" [ 720.133496] env[61991]: _type = "HttpNfcLease" [ 720.133496] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 720.134644] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 720.134644] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cc074f-4742-0d9f-9642-c4cc781292b8" [ 720.134644] env[61991]: _type = "HttpNfcLease" [ 720.134644] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 720.134644] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2da083-96be-4d4f-a551-5ac6b5a942d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.144671] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e1bf6d-31b3-0ea9-4cea-2c95cfefe632/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 720.144671] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e1bf6d-31b3-0ea9-4cea-2c95cfefe632/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 720.295895] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dcd97dc4-cc1d-417e-a084-40c7a7c9c3a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.326269] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521f4dc0-07d2-21de-1378-14bc0c644a13, 'name': SearchDatastore_Task, 'duration_secs': 0.008907} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.332260] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.332549] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.332766] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.389037] env[61991]: DEBUG nova.network.neutron [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updated VIF entry in instance network info cache for port 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.389424] env[61991]: DEBUG nova.network.neutron [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updating instance_info_cache with network_info: [{"id": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "address": "fa:16:3e:8e:7c:67", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fdca45f-c9", "ovs_interfaceid": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.409798] env[61991]: DEBUG nova.compute.utils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.411871] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.412077] env[61991]: DEBUG nova.network.neutron [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 720.446417] env[61991]: DEBUG nova.network.neutron [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updated VIF entry in instance network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.446766] env[61991]: DEBUG nova.network.neutron [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.480220] env[61991]: DEBUG nova.policy [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c5daf21b21b40c9a54bed64a64538bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '822d7e3c678e4defa24bb4d8439a62de', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.496440] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bcad1d-93f3-74e8-c683-447d36eda9e3, 'name': SearchDatastore_Task, 'duration_secs': 0.008455} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.496737] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.496933] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 0631634f-5c58-4ab0-b0dd-71821caf2fc6/0631634f-5c58-4ab0-b0dd-71821caf2fc6.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 720.497225] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.497402] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 720.497605] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e465b3f6-7e5a-4abe-ba07-bec74f91eb99 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.499564] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-041aa82d-b8cb-476e-be2c-7e09548f54e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.507664] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 720.507664] env[61991]: value = "task-1129489" [ 720.507664] env[61991]: _type = "Task" [ 720.507664] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.511765] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.511765] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 720.512774] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8836dd48-2e26-4813-8084-2add983434b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.517926] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129489, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.525544] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 720.525544] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528dc44d-f19d-ad19-71b4-805401019a1b" [ 720.525544] env[61991]: _type = "Task" [ 720.525544] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.535216] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.537445] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528dc44d-f19d-ad19-71b4-805401019a1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.560269] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.560588] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.560753] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.561037] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.561117] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.561252] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.561464] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.561628] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.561804] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.561957] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.562368] env[61991]: DEBUG nova.virt.hardware [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.563292] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2a8c48-33c1-4831-b560-30173b3ecf77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.573818] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d056ebe-6d52-45e5-8af5-be61fa6b223c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.895729] env[61991]: DEBUG oslo_concurrency.lockutils [req-6ae70ebd-bc7e-4628-99c9-27aa00275e66 req-c3b6559e-2bf8-48e5-a324-b638f8ba5d77 service nova] Releasing lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.915910] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.949345] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6565f8f-a60e-40f7-b9e1-28d109d9e20f req-42abd404-da4c-4905-82f8-5df4a348ac8a service nova] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.013722] env[61991]: DEBUG nova.network.neutron [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Successfully created port: dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.026878] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129489, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450988} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.033182] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 0631634f-5c58-4ab0-b0dd-71821caf2fc6/0631634f-5c58-4ab0-b0dd-71821caf2fc6.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.033458] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.036681] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0bcd999-68c1-40f9-99a5-5528a2d8e4d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.044860] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528dc44d-f19d-ad19-71b4-805401019a1b, 'name': SearchDatastore_Task, 'duration_secs': 0.008526} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.047502] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 721.047502] env[61991]: value = "task-1129490" [ 721.047502] env[61991]: _type = "Task" [ 721.047502] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.047502] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beeb9e05-b9a9-4680-ac40-44cb22653069 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.061215] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 721.061215] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fc55cd-9ea5-e678-d535-a6f47e302f7b" [ 721.061215] env[61991]: _type = "Task" [ 721.061215] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.068458] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.075959] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fc55cd-9ea5-e678-d535-a6f47e302f7b, 'name': SearchDatastore_Task, 'duration_secs': 0.009304} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.079028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.079647] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 721.080246] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d8c4d69-a1de-4cdd-86d8-f5870eef18c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.088196] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 721.088196] env[61991]: value = "task-1129491" [ 721.088196] env[61991]: _type = "Task" [ 721.088196] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.095707] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.106952] env[61991]: DEBUG nova.compute.manager [req-b2842993-2c04-49e2-b6ff-51cc16d9b84d req-26a8383a-f536-46f4-a9a7-b217b21b9d6a service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Received event network-vif-plugged-227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.107264] env[61991]: DEBUG oslo_concurrency.lockutils [req-b2842993-2c04-49e2-b6ff-51cc16d9b84d req-26a8383a-f536-46f4-a9a7-b217b21b9d6a service nova] Acquiring lock "30bd843a-9d86-4de1-b011-600e2a93bcb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.107742] env[61991]: DEBUG oslo_concurrency.lockutils [req-b2842993-2c04-49e2-b6ff-51cc16d9b84d req-26a8383a-f536-46f4-a9a7-b217b21b9d6a service nova] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.107914] env[61991]: DEBUG oslo_concurrency.lockutils [req-b2842993-2c04-49e2-b6ff-51cc16d9b84d req-26a8383a-f536-46f4-a9a7-b217b21b9d6a service nova] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.108198] env[61991]: DEBUG nova.compute.manager [req-b2842993-2c04-49e2-b6ff-51cc16d9b84d req-26a8383a-f536-46f4-a9a7-b217b21b9d6a service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] No waiting events found dispatching network-vif-plugged-227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 721.108378] env[61991]: WARNING nova.compute.manager [req-b2842993-2c04-49e2-b6ff-51cc16d9b84d req-26a8383a-f536-46f4-a9a7-b217b21b9d6a service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Received unexpected event network-vif-plugged-227fb866-9358-46ef-a164-201f66f64060 for instance with vm_state building and task_state spawning. [ 721.180565] env[61991]: DEBUG nova.network.neutron [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Successfully updated port: 227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 721.380569] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7906846d-1484-4385-91eb-e1956fd6f422 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.389048] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88277d9-eefb-47bf-9dea-815ee406444d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.430603] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8f5de0-44f2-4dec-8502-b8930c65ca72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.439750] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409ba30c-e7f9-426e-bf49-c31f32e81ce0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.456457] env[61991]: DEBUG nova.compute.provider_tree [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.558717] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076864} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.558990] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.559853] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c802654-e688-45ab-8d8a-f9f0a80f2343 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.586731] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 0631634f-5c58-4ab0-b0dd-71821caf2fc6/0631634f-5c58-4ab0-b0dd-71821caf2fc6.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.587274] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b07c6fce-6e90-47a7-a468-448f5de28c69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.610704] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129491, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437449} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.612601] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.613301] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.613301] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 721.613301] env[61991]: value = "task-1129492" [ 721.613301] env[61991]: _type = "Task" [ 721.613301] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.613489] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66bb6040-47de-4b2f-a9cb-7fe1ca54781b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.624080] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129492, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.625399] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 721.625399] env[61991]: value = "task-1129493" [ 721.625399] env[61991]: _type = "Task" [ 721.625399] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.633380] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129493, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.684500] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "refresh_cache-30bd843a-9d86-4de1-b011-600e2a93bcb3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.684697] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquired lock "refresh_cache-30bd843a-9d86-4de1-b011-600e2a93bcb3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.684868] env[61991]: DEBUG nova.network.neutron [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.936625] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.960866] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.961802] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.961802] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.961802] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.961802] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.961802] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.962066] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.962106] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.962278] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.962442] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.962615] env[61991]: DEBUG nova.virt.hardware [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.963604] env[61991]: DEBUG nova.scheduler.client.report [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.967108] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078475a2-bfb4-4bf3-bf05-fa33263759f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.975836] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e90a33-da0b-4ec2-9089-b2815b888c4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.126792] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129492, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.134517] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129493, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127762} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.134761] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.135528] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2690347-fe4d-4528-8a87-d8861274c468 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.158476] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.158754] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4252ece6-1cc1-4861-95e8-ac6541763d8d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.178455] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 722.178455] env[61991]: value = "task-1129494" [ 722.178455] env[61991]: _type = "Task" [ 722.178455] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.188924] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129494, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.231474] env[61991]: DEBUG nova.network.neutron [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.417420] env[61991]: DEBUG nova.network.neutron [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Updating instance_info_cache with network_info: [{"id": "227fb866-9358-46ef-a164-201f66f64060", "address": "fa:16:3e:88:27:b0", "network": {"id": "8fcbd719-6150-4a27-a644-f3aa42b99d88", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1697011133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be45d020174c488b9a55e45bed5ac324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227fb866-93", "ovs_interfaceid": "227fb866-9358-46ef-a164-201f66f64060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.471171] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.473636] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.475657] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.498s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.475992] env[61991]: DEBUG nova.objects.instance [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 722.628403] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129492, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.677998] env[61991]: DEBUG nova.compute.manager [req-beda2cd8-5d7d-4fc8-a381-e3d0ab2fcbb3 req-e8847fda-d7d2-4582-a94d-27e0b362a38d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Received event network-vif-plugged-dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.678169] env[61991]: DEBUG oslo_concurrency.lockutils [req-beda2cd8-5d7d-4fc8-a381-e3d0ab2fcbb3 req-e8847fda-d7d2-4582-a94d-27e0b362a38d service nova] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.678393] env[61991]: DEBUG oslo_concurrency.lockutils [req-beda2cd8-5d7d-4fc8-a381-e3d0ab2fcbb3 req-e8847fda-d7d2-4582-a94d-27e0b362a38d service nova] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.678564] env[61991]: DEBUG oslo_concurrency.lockutils [req-beda2cd8-5d7d-4fc8-a381-e3d0ab2fcbb3 req-e8847fda-d7d2-4582-a94d-27e0b362a38d service nova] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.678736] env[61991]: DEBUG nova.compute.manager [req-beda2cd8-5d7d-4fc8-a381-e3d0ab2fcbb3 req-e8847fda-d7d2-4582-a94d-27e0b362a38d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] No waiting events found dispatching network-vif-plugged-dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 722.678957] env[61991]: WARNING nova.compute.manager [req-beda2cd8-5d7d-4fc8-a381-e3d0ab2fcbb3 req-e8847fda-d7d2-4582-a94d-27e0b362a38d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Received unexpected event network-vif-plugged-dff90314-9bf7-4448-980c-047fc0a617ec for instance with vm_state building and task_state spawning. [ 722.689187] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129494, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.842026] env[61991]: DEBUG nova.network.neutron [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Successfully updated port: dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 722.924941] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Releasing lock "refresh_cache-30bd843a-9d86-4de1-b011-600e2a93bcb3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.924941] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Instance network_info: |[{"id": "227fb866-9358-46ef-a164-201f66f64060", "address": "fa:16:3e:88:27:b0", "network": {"id": "8fcbd719-6150-4a27-a644-f3aa42b99d88", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1697011133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be45d020174c488b9a55e45bed5ac324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227fb866-93", "ovs_interfaceid": "227fb866-9358-46ef-a164-201f66f64060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 722.925223] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:27:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '227fb866-9358-46ef-a164-201f66f64060', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.933617] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Creating folder: Project (be45d020174c488b9a55e45bed5ac324). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.933617] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-416ff692-5c0b-4028-a47a-f99677854fb1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.943603] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Created folder: Project (be45d020174c488b9a55e45bed5ac324) in parent group-v246753. [ 722.944483] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Creating folder: Instances. Parent ref: group-v246819. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.944483] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc1f3725-237b-403f-96a3-8d75c2e9374b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.954514] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Created folder: Instances in parent group-v246819. [ 722.954795] env[61991]: DEBUG oslo.service.loopingcall [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.955022] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.955253] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fc36e93-f1f5-47fa-a1e1-b87404921369 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.975418] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.975418] env[61991]: value = "task-1129497" [ 722.975418] env[61991]: _type = "Task" [ 722.975418] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.980368] env[61991]: DEBUG nova.compute.utils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.988466] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.988466] env[61991]: DEBUG nova.network.neutron [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.991029] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129497, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.046530] env[61991]: DEBUG nova.policy [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.127683] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129492, 'name': ReconfigVM_Task, 'duration_secs': 1.278886} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.128042] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 0631634f-5c58-4ab0-b0dd-71821caf2fc6/0631634f-5c58-4ab0-b0dd-71821caf2fc6.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.128746] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62b16105-1ad0-4d86-b7bf-f4de7dba3912 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.135739] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 723.135739] env[61991]: value = "task-1129498" [ 723.135739] env[61991]: _type = "Task" [ 723.135739] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.141206] env[61991]: DEBUG nova.compute.manager [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Received event network-changed-227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.141581] env[61991]: DEBUG nova.compute.manager [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Refreshing instance network info cache due to event network-changed-227fb866-9358-46ef-a164-201f66f64060. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 723.143480] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] Acquiring lock "refresh_cache-30bd843a-9d86-4de1-b011-600e2a93bcb3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.143681] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] Acquired lock "refresh_cache-30bd843a-9d86-4de1-b011-600e2a93bcb3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.143845] env[61991]: DEBUG nova.network.neutron [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Refreshing network info cache for port 227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.150692] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129498, 'name': Rename_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.192033] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129494, 'name': ReconfigVM_Task, 'duration_secs': 0.514942} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.192033] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Reconfigured VM instance instance-00000031 to attach disk [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.192033] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c3983d1-fd48-4585-80db-d93d58f92a7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.198750] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 723.198750] env[61991]: value = "task-1129499" [ 723.198750] env[61991]: _type = "Task" [ 723.198750] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.214463] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129499, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.349066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.349066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.349066] env[61991]: DEBUG nova.network.neutron [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.393498] env[61991]: DEBUG nova.network.neutron [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Successfully created port: ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.485840] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129497, 'name': CreateVM_Task, 'duration_secs': 0.343064} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.486421] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.487212] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.487462] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.487724] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 723.488419] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2149280f-f463-45b8-bb96-00bd213d22f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.490863] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.495025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-401cf153-c356-451e-86bf-f78e8b168711 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.497338] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.238s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.502289] env[61991]: INFO nova.compute.claims [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.506163] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 723.506163] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b435a9-4e62-e6ad-8767-f79d588d7d49" [ 723.506163] env[61991]: _type = "Task" [ 723.506163] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.518443] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b435a9-4e62-e6ad-8767-f79d588d7d49, 'name': SearchDatastore_Task, 'duration_secs': 0.009643} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.518752] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.518986] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.519243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.519395] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.519575] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.519839] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f8eac4b-494c-43ec-bdbf-30ccacb3c614 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.528904] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.529116] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.529879] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a9c259b-d292-4461-9099-13a8c382ad18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.536231] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 723.536231] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5241dd90-d1c7-a592-0c39-7ff54f384536" [ 723.536231] env[61991]: _type = "Task" [ 723.536231] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.548021] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5241dd90-d1c7-a592-0c39-7ff54f384536, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.650340] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129498, 'name': Rename_Task, 'duration_secs': 0.145694} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.650577] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.651639] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9430c6c0-857f-41e8-936c-7f63b6b2e6c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.657696] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 723.657696] env[61991]: value = "task-1129500" [ 723.657696] env[61991]: _type = "Task" [ 723.657696] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.669136] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129500, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.709958] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129499, 'name': Rename_Task, 'duration_secs': 0.161515} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.712931] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.712931] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31b28f8f-1957-4dc8-8ac8-7ebe743c24f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.718507] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 723.718507] env[61991]: value = "task-1129501" [ 723.718507] env[61991]: _type = "Task" [ 723.718507] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.727401] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129501, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.910402] env[61991]: DEBUG nova.network.neutron [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.002046] env[61991]: DEBUG nova.network.neutron [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Updated VIF entry in instance network info cache for port 227fb866-9358-46ef-a164-201f66f64060. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 724.002496] env[61991]: DEBUG nova.network.neutron [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Updating instance_info_cache with network_info: [{"id": "227fb866-9358-46ef-a164-201f66f64060", "address": "fa:16:3e:88:27:b0", "network": {"id": "8fcbd719-6150-4a27-a644-f3aa42b99d88", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1697011133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "be45d020174c488b9a55e45bed5ac324", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227fb866-93", "ovs_interfaceid": "227fb866-9358-46ef-a164-201f66f64060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.053359] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5241dd90-d1c7-a592-0c39-7ff54f384536, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.054244] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7e3a1f8-c430-41da-b56c-fbe23bf93699 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.060637] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 724.060637] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cbb0b2-93cd-2002-8277-16d9a939ebf4" [ 724.060637] env[61991]: _type = "Task" [ 724.060637] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.074370] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cbb0b2-93cd-2002-8277-16d9a939ebf4, 'name': SearchDatastore_Task, 'duration_secs': 0.01012} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.074725] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.075081] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 30bd843a-9d86-4de1-b011-600e2a93bcb3/30bd843a-9d86-4de1-b011-600e2a93bcb3.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.075372] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57fe7efb-e0c6-43f5-8533-b379819abcf9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.084394] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 724.084394] env[61991]: value = "task-1129502" [ 724.084394] env[61991]: _type = "Task" [ 724.084394] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.094035] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.168587] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129500, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.200302] env[61991]: DEBUG nova.network.neutron [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.229934] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129501, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.505749] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.509330] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d50d244-c2f7-410b-ab20-d76dcb5d3db1 req-c1564067-a720-43f1-bfd3-cfc0e48d65b0 service nova] Releasing lock "refresh_cache-30bd843a-9d86-4de1-b011-600e2a93bcb3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.534479] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.534877] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.535151] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.535483] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.535710] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.535953] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.536311] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.536582] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.537471] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.537471] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.537471] env[61991]: DEBUG nova.virt.hardware [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.538830] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4539f4-f674-4caf-93c8-42bf3e359c89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.556251] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c1d6e0-8cbd-456d-865a-1b5a654bd090 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.598941] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129502, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.670097] env[61991]: DEBUG oslo_vmware.api [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129500, 'name': PowerOnVM_Task, 'duration_secs': 0.548338} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.673540] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.673775] env[61991]: INFO nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Took 12.16 seconds to spawn the instance on the hypervisor. [ 724.674025] env[61991]: DEBUG nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 724.675059] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fa93cd-f436-49ff-b5f7-5a11a10b33d3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.703931] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.703931] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Instance network_info: |[{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 724.703931] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:2d:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dff90314-9bf7-4448-980c-047fc0a617ec', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.712246] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating folder: Project (822d7e3c678e4defa24bb4d8439a62de). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 724.712246] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3587a3e2-c52c-4f30-bf74-ce5c59ebe525 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.725466] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created folder: Project (822d7e3c678e4defa24bb4d8439a62de) in parent group-v246753. [ 724.725673] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating folder: Instances. Parent ref: group-v246822. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 724.725935] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-923d563e-f269-4a4f-a327-b9793f009fef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.733171] env[61991]: DEBUG oslo_vmware.api [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129501, 'name': PowerOnVM_Task, 'duration_secs': 0.646142} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.733914] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.733914] env[61991]: INFO nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Took 7.72 seconds to spawn the instance on the hypervisor. [ 724.733914] env[61991]: DEBUG nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 724.738216] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8fd226a-b15e-4e6f-af2e-f1d5b2e666a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.741204] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created folder: Instances in parent group-v246822. [ 724.741438] env[61991]: DEBUG oslo.service.loopingcall [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.741885] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 724.742451] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06f92a43-23bb-4c70-ba73-bb891f42ca98 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.770726] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.770726] env[61991]: value = "task-1129505" [ 724.770726] env[61991]: _type = "Task" [ 724.770726] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.782318] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129505, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.919941] env[61991]: DEBUG nova.compute.manager [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Received event network-changed-dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.920213] env[61991]: DEBUG nova.compute.manager [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Refreshing instance network info cache due to event network-changed-dff90314-9bf7-4448-980c-047fc0a617ec. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.920617] env[61991]: DEBUG oslo_concurrency.lockutils [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.920808] env[61991]: DEBUG oslo_concurrency.lockutils [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.920987] env[61991]: DEBUG nova.network.neutron [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Refreshing network info cache for port dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.000967] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa5c8f0-b2fa-41fd-beec-8f07b1c7a964 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.009024] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ae992a-cfc4-4cfc-bddc-fda01d540ade {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.039840] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499941ec-c75c-4e44-a51a-dd762d024ac7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.047604] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2caf90-af19-475a-bca2-fadd8cbad877 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.061622] env[61991]: DEBUG nova.compute.provider_tree [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.096044] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54719} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.096044] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 30bd843a-9d86-4de1-b011-600e2a93bcb3/30bd843a-9d86-4de1-b011-600e2a93bcb3.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 725.096044] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 725.096044] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7d9f400-e77c-4f45-a7de-6d7e263152b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.103227] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 725.103227] env[61991]: value = "task-1129506" [ 725.103227] env[61991]: _type = "Task" [ 725.103227] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.112127] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129506, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.194815] env[61991]: INFO nova.compute.manager [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Took 40.87 seconds to build instance. [ 725.252645] env[61991]: DEBUG nova.network.neutron [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Successfully updated port: ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 725.278943] env[61991]: INFO nova.compute.manager [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Took 33.09 seconds to build instance. [ 725.283550] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129505, 'name': CreateVM_Task, 'duration_secs': 0.444173} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.283958] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 725.284586] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.284752] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.285113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 725.285534] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b996627f-363e-4870-8839-41859bdcf825 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.290687] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 725.290687] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52859711-99cb-181a-240e-eb23dd0e6307" [ 725.290687] env[61991]: _type = "Task" [ 725.290687] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.299564] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52859711-99cb-181a-240e-eb23dd0e6307, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.565954] env[61991]: DEBUG nova.scheduler.client.report [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.614885] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129506, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122991} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.615189] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.616031] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b62583d-1df4-4c22-beff-e19ceab18363 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.640936] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 30bd843a-9d86-4de1-b011-600e2a93bcb3/30bd843a-9d86-4de1-b011-600e2a93bcb3.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.644058] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c33b9b4a-6906-413e-81e2-eec344f11585 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.667196] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 725.667196] env[61991]: value = "task-1129507" [ 725.667196] env[61991]: _type = "Task" [ 725.667196] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.679116] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129507, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.696262] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0291952a-2a63-4219-978e-09bd9ebfb8b5 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.228s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.720609] env[61991]: DEBUG nova.network.neutron [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updated VIF entry in instance network info cache for port dff90314-9bf7-4448-980c-047fc0a617ec. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 725.720825] env[61991]: DEBUG nova.network.neutron [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.755753] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.755981] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.756258] env[61991]: DEBUG nova.network.neutron [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.784494] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6041b36-fe02-457e-a0fa-7d390ca04952 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.995s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.801955] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52859711-99cb-181a-240e-eb23dd0e6307, 'name': SearchDatastore_Task, 'duration_secs': 0.019726} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.802325] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.802641] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.802827] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.803072] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.803200] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.803418] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d50fdc3e-d2c3-44e0-9372-c95e6d3f84dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.812799] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.812799] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 725.813589] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24c1933d-0661-4b4c-b229-624cd5bb7f2f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.819888] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 725.819888] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f4986e-0a1e-f640-72a6-122fd992c3fb" [ 725.819888] env[61991]: _type = "Task" [ 725.819888] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.827606] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f4986e-0a1e-f640-72a6-122fd992c3fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.017576] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.017857] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.018090] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.018295] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.018476] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.020831] env[61991]: INFO nova.compute.manager [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Terminating instance [ 726.022806] env[61991]: DEBUG nova.compute.manager [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.023010] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.023873] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b33c375-4bda-4f00-8a98-0b7100bc165f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.034531] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 726.034801] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e669c592-6a6e-4d89-ac10-92451987c3e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.041205] env[61991]: DEBUG oslo_vmware.api [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 726.041205] env[61991]: value = "task-1129508" [ 726.041205] env[61991]: _type = "Task" [ 726.041205] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.049445] env[61991]: DEBUG oslo_vmware.api [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.071066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.072159] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.074817] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.389s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.075170] env[61991]: DEBUG nova.objects.instance [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lazy-loading 'resources' on Instance uuid 812d3491-0598-4171-b905-5304ebf6ae37 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.181194] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129507, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.203303] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.224224] env[61991]: DEBUG oslo_concurrency.lockutils [req-303ccabb-8b19-4c81-a7ff-28de7e5906b8 req-c50c5add-5ecc-47d2-a37d-49fc1c70f89d service nova] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.287701] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.291624] env[61991]: DEBUG nova.network.neutron [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.333361] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f4986e-0a1e-f640-72a6-122fd992c3fb, 'name': SearchDatastore_Task, 'duration_secs': 0.025789} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.334441] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea833355-8b7d-4fdb-82e9-61e5946246df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.347344] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 726.347344] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52473c06-1066-f457-49f5-5ebd0399a4c3" [ 726.347344] env[61991]: _type = "Task" [ 726.347344] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.357322] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52473c06-1066-f457-49f5-5ebd0399a4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.009033} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.357666] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.357968] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556/3bcde671-5702-4b8a-8881-88eb7dfd0556.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 726.358281] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8cbdf47-21d4-4185-8756-39d3a8c2fd9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.365931] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 726.365931] env[61991]: value = "task-1129509" [ 726.365931] env[61991]: _type = "Task" [ 726.365931] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.375223] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129509, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.460630] env[61991]: DEBUG nova.compute.manager [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.460787] env[61991]: DEBUG nova.compute.manager [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing instance network info cache due to event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.462155] env[61991]: DEBUG oslo_concurrency.lockutils [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.462155] env[61991]: DEBUG oslo_concurrency.lockutils [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.462371] env[61991]: DEBUG nova.network.neutron [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.502801] env[61991]: DEBUG nova.network.neutron [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.555964] env[61991]: DEBUG oslo_vmware.api [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129508, 'name': PowerOffVM_Task, 'duration_secs': 0.196846} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.556747] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 726.556956] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 726.557249] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9f0b342-5324-4c99-aa58-3d18f0f4ba06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.579514] env[61991]: DEBUG nova.compute.utils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.586055] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.586055] env[61991]: DEBUG nova.network.neutron [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.646524] env[61991]: DEBUG nova.policy [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e6810d3aced4a44bbfe9ba49270bc35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a18b2a73f7643e1b89c9660a770e9da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.692022] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129507, 'name': ReconfigVM_Task, 'duration_secs': 0.529221} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.693796] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 30bd843a-9d86-4de1-b011-600e2a93bcb3/30bd843a-9d86-4de1-b011-600e2a93bcb3.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.695071] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be255715-8ffb-4c56-9c59-fafa869729cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.705149] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 726.705149] env[61991]: value = "task-1129511" [ 726.705149] env[61991]: _type = "Task" [ 726.705149] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.713398] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 726.713896] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 726.714240] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Deleting the datastore file [datastore1] 0631634f-5c58-4ab0-b0dd-71821caf2fc6 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 726.722638] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3ef958e-2a17-4187-927a-6eeb82cf377b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.729362] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129511, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.738484] env[61991]: DEBUG oslo_vmware.api [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for the task: (returnval){ [ 726.738484] env[61991]: value = "task-1129512" [ 726.738484] env[61991]: _type = "Task" [ 726.738484] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.746693] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.754751] env[61991]: DEBUG oslo_vmware.api [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129512, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.828308] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.880312] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129509, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466886} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.880701] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556/3bcde671-5702-4b8a-8881-88eb7dfd0556.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 726.880852] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.881323] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4b9034c-c0fb-45a8-a2db-834c3431f8fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.888863] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 726.888863] env[61991]: value = "task-1129513" [ 726.888863] env[61991]: _type = "Task" [ 726.888863] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.901635] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129513, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.005967] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.006545] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Instance network_info: |[{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 727.009783] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:32:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3f695b6-65bc-45cc-a61d-3c38a14e5c0c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea66ec35-3e8b-40d2-baa8-a6d811b6e7be', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.021021] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Creating folder: Project (d58eaa6ccc4f46e3a7606f55605d8e15). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.022358] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-848cd462-0fef-49c7-bf94-9e1f98d89f03 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.033639] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Created folder: Project (d58eaa6ccc4f46e3a7606f55605d8e15) in parent group-v246753. [ 727.034117] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Creating folder: Instances. Parent ref: group-v246825. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.038114] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5bb64d0-186d-4c17-bf99-6a50afb3cf35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.048132] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Created folder: Instances in parent group-v246825. [ 727.048431] env[61991]: DEBUG oslo.service.loopingcall [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.048753] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.048911] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71e79e3f-a305-4834-b23a-7ac35ba7098c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.075567] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.075567] env[61991]: value = "task-1129516" [ 727.075567] env[61991]: _type = "Task" [ 727.075567] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.086319] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.092887] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129516, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.120631] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc91f1b-7778-4133-aece-7dc560d5c492 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.126303] env[61991]: DEBUG nova.compute.manager [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-vif-plugged-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.126547] env[61991]: DEBUG oslo_concurrency.lockutils [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.127216] env[61991]: DEBUG oslo_concurrency.lockutils [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.127403] env[61991]: DEBUG oslo_concurrency.lockutils [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.128158] env[61991]: DEBUG nova.compute.manager [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] No waiting events found dispatching network-vif-plugged-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 727.128158] env[61991]: WARNING nova.compute.manager [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received unexpected event network-vif-plugged-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be for instance with vm_state building and task_state spawning. [ 727.128158] env[61991]: DEBUG nova.compute.manager [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-changed-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.128158] env[61991]: DEBUG nova.compute.manager [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing instance network info cache due to event network-changed-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.129293] env[61991]: DEBUG oslo_concurrency.lockutils [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.129293] env[61991]: DEBUG oslo_concurrency.lockutils [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.129293] env[61991]: DEBUG nova.network.neutron [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing network info cache for port ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.136542] env[61991]: DEBUG nova.network.neutron [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Successfully created port: 8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.139640] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c4b1be-4b9b-48eb-b604-f8a795f4994b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.178028] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a11780-4362-464a-b49d-17a3b2d80fb9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.189357] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd8252e-073f-4da6-9423-d9104ef87469 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.206976] env[61991]: DEBUG nova.compute.provider_tree [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.218206] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129511, 'name': Rename_Task, 'duration_secs': 0.209023} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.221378] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 727.222394] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97aa0c9f-545a-49cb-a92f-634fd7f0257a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.230710] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 727.230710] env[61991]: value = "task-1129517" [ 727.230710] env[61991]: _type = "Task" [ 727.230710] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.239183] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.247754] env[61991]: DEBUG oslo_vmware.api [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Task: {'id': task-1129512, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354412} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.247992] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 727.248293] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 727.248612] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.248719] env[61991]: INFO nova.compute.manager [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Took 1.23 seconds to destroy the instance on the hypervisor. [ 727.248933] env[61991]: DEBUG oslo.service.loopingcall [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.249143] env[61991]: DEBUG nova.compute.manager [-] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.249243] env[61991]: DEBUG nova.network.neutron [-] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.375624] env[61991]: DEBUG nova.network.neutron [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updated VIF entry in instance network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 727.376051] env[61991]: DEBUG nova.network.neutron [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.400125] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129513, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.169368} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.403537] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 727.404334] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e432481c-1f63-4c83-a1f4-51fdfe26d863 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.436770] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556/3bcde671-5702-4b8a-8881-88eb7dfd0556.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.438144] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0794cc18-6ff8-4e2e-adf1-29732a240503 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.463287] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 727.463287] env[61991]: value = "task-1129518" [ 727.463287] env[61991]: _type = "Task" [ 727.463287] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.474449] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129518, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.587693] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129516, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.710178] env[61991]: DEBUG nova.scheduler.client.report [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.743701] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129517, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.879520] env[61991]: DEBUG oslo_concurrency.lockutils [req-9723f79b-4a76-4594-9fcb-b6f93c2a1bfc req-f3de22db-5c1b-459d-acc3-82fb0ea60bf6 service nova] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.975723] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129518, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.982169] env[61991]: DEBUG nova.network.neutron [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updated VIF entry in instance network info cache for port ea66ec35-3e8b-40d2-baa8-a6d811b6e7be. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 727.982515] env[61991]: DEBUG nova.network.neutron [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.088616] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129516, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.101137] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.123743] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.124054] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.124216] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.124412] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.124598] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.124708] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.124917] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.125138] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.125362] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.125553] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.125755] env[61991]: DEBUG nova.virt.hardware [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.127190] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe823909-da16-4a9d-8e2c-944367cfb75a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.135296] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e799023-3521-48ec-b3f4-b6b944a524d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.218917] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.144s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.221426] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 22.578s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.244026] env[61991]: DEBUG oslo_vmware.api [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129517, 'name': PowerOnVM_Task, 'duration_secs': 0.854432} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.245414] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 728.245687] env[61991]: INFO nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Took 7.71 seconds to spawn the instance on the hypervisor. [ 728.246143] env[61991]: DEBUG nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 728.247326] env[61991]: INFO nova.scheduler.client.report [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleted allocations for instance 812d3491-0598-4171-b905-5304ebf6ae37 [ 728.249030] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1192f3-1bbe-4d29-847d-95767882ed14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.406728] env[61991]: DEBUG nova.network.neutron [-] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.475165] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129518, 'name': ReconfigVM_Task, 'duration_secs': 0.943504} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.475165] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556/3bcde671-5702-4b8a-8881-88eb7dfd0556.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.475739] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57e710eb-0d04-483b-9e16-78f3a3adaff2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.482259] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 728.482259] env[61991]: value = "task-1129519" [ 728.482259] env[61991]: _type = "Task" [ 728.482259] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.486338] env[61991]: DEBUG oslo_concurrency.lockutils [req-27ec97c7-4fae-4e4b-8f05-0bc393a0138c req-627e63dd-88ba-40a8-89a8-c4a2a2490429 service nova] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.490262] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129519, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.492342] env[61991]: DEBUG nova.compute.manager [req-daa7c0fb-b8da-4586-979e-816de927d1a0 req-d3f60860-1c53-469d-abb0-682486f7e10f service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received event network-vif-deleted-15c923b4-4e91-4367-9346-cb8b6e51c275 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.492548] env[61991]: DEBUG nova.compute.manager [req-daa7c0fb-b8da-4586-979e-816de927d1a0 req-d3f60860-1c53-469d-abb0-682486f7e10f service nova] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Received event network-vif-deleted-4c68db77-3ed0-4069-ab22-78bb0f625e1c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.587620] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129516, 'name': CreateVM_Task, 'duration_secs': 1.279449} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.587846] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 728.588658] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.588878] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.589274] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 728.589588] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc32e814-1ecb-46d0-8d14-89b75a638cb7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.594457] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 728.594457] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5225bf20-aa13-5b36-6652-a6fffac244be" [ 728.594457] env[61991]: _type = "Task" [ 728.594457] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.602884] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5225bf20-aa13-5b36-6652-a6fffac244be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.727598] env[61991]: INFO nova.compute.claims [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.758516] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eee9f4b2-3d03-4083-8341-3106e1243d2e tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "812d3491-0598-4171-b905-5304ebf6ae37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.056s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.767794] env[61991]: INFO nova.compute.manager [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Took 36.52 seconds to build instance. [ 728.878609] env[61991]: DEBUG nova.compute.manager [req-0501b29b-1de6-4f2d-b1bd-a04fac6d799c req-94bd6580-85f6-4b47-9fbf-80793c2b2cd8 service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Received event network-vif-plugged-8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.878832] env[61991]: DEBUG oslo_concurrency.lockutils [req-0501b29b-1de6-4f2d-b1bd-a04fac6d799c req-94bd6580-85f6-4b47-9fbf-80793c2b2cd8 service nova] Acquiring lock "78631e83-e574-4a90-b062-176b2d12d681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.879123] env[61991]: DEBUG oslo_concurrency.lockutils [req-0501b29b-1de6-4f2d-b1bd-a04fac6d799c req-94bd6580-85f6-4b47-9fbf-80793c2b2cd8 service nova] Lock "78631e83-e574-4a90-b062-176b2d12d681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.879310] env[61991]: DEBUG oslo_concurrency.lockutils [req-0501b29b-1de6-4f2d-b1bd-a04fac6d799c req-94bd6580-85f6-4b47-9fbf-80793c2b2cd8 service nova] Lock "78631e83-e574-4a90-b062-176b2d12d681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.879485] env[61991]: DEBUG nova.compute.manager [req-0501b29b-1de6-4f2d-b1bd-a04fac6d799c req-94bd6580-85f6-4b47-9fbf-80793c2b2cd8 service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] No waiting events found dispatching network-vif-plugged-8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 728.879654] env[61991]: WARNING nova.compute.manager [req-0501b29b-1de6-4f2d-b1bd-a04fac6d799c req-94bd6580-85f6-4b47-9fbf-80793c2b2cd8 service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Received unexpected event network-vif-plugged-8aa50c96-062e-4d6c-9aa0-990a50838237 for instance with vm_state building and task_state spawning. [ 728.909500] env[61991]: INFO nova.compute.manager [-] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Took 1.66 seconds to deallocate network for instance. [ 728.992420] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129519, 'name': Rename_Task, 'duration_secs': 0.162322} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.992751] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 728.993037] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb63683c-abeb-4898-84d3-51d8859a90a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.999705] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 728.999705] env[61991]: value = "task-1129520" [ 728.999705] env[61991]: _type = "Task" [ 728.999705] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.007858] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129520, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.107076] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5225bf20-aa13-5b36-6652-a6fffac244be, 'name': SearchDatastore_Task, 'duration_secs': 0.012247} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.107451] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.107701] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.107939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.108117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.108332] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.108598] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d261c06a-d0f2-48ca-be47-88ee6cd4e662 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.119144] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.119525] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 729.123041] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a955e859-6de3-4698-b4ae-37f8045b13f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.127929] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 729.127929] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ced235-3b1a-acb6-51bb-28e05fb5eea5" [ 729.127929] env[61991]: _type = "Task" [ 729.127929] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.136629] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ced235-3b1a-acb6-51bb-28e05fb5eea5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.233887] env[61991]: INFO nova.compute.resource_tracker [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating resource usage from migration 528ff66f-4caf-4189-ac3a-ace614777b67 [ 729.273518] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab175fe8-3482-49ee-83e2-82965e9e821a tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.529s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.390332] env[61991]: DEBUG nova.network.neutron [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Successfully updated port: 8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 729.416089] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.443977] env[61991]: DEBUG nova.compute.manager [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Received event network-changed-8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.444110] env[61991]: DEBUG nova.compute.manager [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Refreshing instance network info cache due to event network-changed-8aa50c96-062e-4d6c-9aa0-990a50838237. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.444396] env[61991]: DEBUG oslo_concurrency.lockutils [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] Acquiring lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.444525] env[61991]: DEBUG oslo_concurrency.lockutils [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] Acquired lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.444590] env[61991]: DEBUG nova.network.neutron [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Refreshing network info cache for port 8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.510577] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129520, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.534723] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e1bf6d-31b3-0ea9-4cea-2c95cfefe632/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 729.535698] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e777e8f5-3c74-42a0-8213-4522caa8067f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.543872] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e1bf6d-31b3-0ea9-4cea-2c95cfefe632/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 729.544100] env[61991]: ERROR oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e1bf6d-31b3-0ea9-4cea-2c95cfefe632/disk-0.vmdk due to incomplete transfer. [ 729.544344] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9d050e19-6b1e-410c-a13b-b08c2b8a1707 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.552847] env[61991]: DEBUG oslo_vmware.rw_handles [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e1bf6d-31b3-0ea9-4cea-2c95cfefe632/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 729.552847] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Uploaded image e4b0de21-26e3-4d65-834b-7a893af37f7a to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 729.555374] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 729.555629] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b754909c-4c7b-440a-87ff-860dad5edfe4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.561246] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 729.561246] env[61991]: value = "task-1129521" [ 729.561246] env[61991]: _type = "Task" [ 729.561246] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.574039] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129521, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.579768] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ec6c8d-2df8-4736-9bfb-9265781352c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.587947] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3197d63-3ccf-4f1b-8352-f56e2a31c1d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.362538] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "30bd843a-9d86-4de1-b011-600e2a93bcb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.362823] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.362980] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "30bd843a-9d86-4de1-b011-600e2a93bcb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.363181] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.363351] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.365033] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.367745] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.370097] env[61991]: INFO nova.compute.manager [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Terminating instance [ 730.381340] env[61991]: DEBUG nova.compute.manager [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 730.382231] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.383269] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81e707f-708d-4584-a0e8-e4457d5fab52 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.415194] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329fd3d3-5953-4220-823a-ef58a1ae9b42 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.425837] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ced235-3b1a-acb6-51bb-28e05fb5eea5, 'name': SearchDatastore_Task, 'duration_secs': 0.017772} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.426148] env[61991]: DEBUG oslo_vmware.api [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129520, 'name': PowerOnVM_Task, 'duration_secs': 1.002054} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.427904] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 730.428140] env[61991]: INFO nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 8.49 seconds to spawn the instance on the hypervisor. [ 730.428330] env[61991]: DEBUG nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 730.432069] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e801a718-6184-4139-84a9-bd128fda06a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.434450] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbe6c61-3ddd-4c43-ba19-6d694e8c5c62 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.437107] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129521, 'name': Destroy_Task, 'duration_secs': 0.328236} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.441290] env[61991]: DEBUG nova.network.neutron [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.443325] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Destroyed the VM [ 730.443634] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 730.443969] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 730.444814] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c917b791-88ea-443c-a14b-e8e7b6b288cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.447374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fff0c1-864e-4650-8e69-5f6592e92c9a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.451510] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-139d7e7e-4649-4072-83af-47ffad617d77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.455740] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 730.455740] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c63b35-2c16-4a17-f9e1-ed91d5885a9c" [ 730.455740] env[61991]: _type = "Task" [ 730.455740] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.474117] env[61991]: DEBUG nova.compute.provider_tree [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.475402] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 730.475402] env[61991]: value = "task-1129522" [ 730.475402] env[61991]: _type = "Task" [ 730.475402] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.477329] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 730.477329] env[61991]: value = "task-1129523" [ 730.477329] env[61991]: _type = "Task" [ 730.477329] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.487071] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c63b35-2c16-4a17-f9e1-ed91d5885a9c, 'name': SearchDatastore_Task, 'duration_secs': 0.023439} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.488320] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.488619] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 80d05278-b516-4408-94b0-11bc93500b5c/80d05278-b516-4408-94b0-11bc93500b5c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 730.489524] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-319e07f7-b054-4994-a187-1aa15e2f5b3e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.498151] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129523, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.498447] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129522, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.502899] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 730.502899] env[61991]: value = "task-1129524" [ 730.502899] env[61991]: _type = "Task" [ 730.502899] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.511361] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.575028] env[61991]: DEBUG nova.network.neutron [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.898610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.971022] env[61991]: INFO nova.compute.manager [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 36.08 seconds to build instance. [ 730.978362] env[61991]: DEBUG nova.scheduler.client.report [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.001354] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129522, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.001621] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129523, 'name': PowerOffVM_Task, 'duration_secs': 0.265271} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.001869] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 731.002042] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 731.002570] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3717a0c-c46a-42c0-ad33-11a99102fe75 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.014790] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129524, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.028488] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.028752] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.077957] env[61991]: DEBUG oslo_concurrency.lockutils [req-7a8f3a11-4223-460b-97c4-c2db15563d43 req-dbf9b911-16d2-4096-9114-c0401cbd7ced service nova] Releasing lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.078536] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.078630] env[61991]: DEBUG nova.network.neutron [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.089820] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 731.089820] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 731.089820] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Deleting the datastore file [datastore1] 30bd843a-9d86-4de1-b011-600e2a93bcb3 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 731.090672] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c38fa8d7-336b-4a01-9896-8ddf5c098cea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.097691] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for the task: (returnval){ [ 731.097691] env[61991]: value = "task-1129526" [ 731.097691] env[61991]: _type = "Task" [ 731.097691] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.106275] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.460856] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "3e62f219-7fec-470f-9d0a-e292d62e076c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.461193] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.471971] env[61991]: DEBUG oslo_concurrency.lockutils [None req-812ae1a2-003b-4188-a1bf-f7102d958525 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.894s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.493967] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.272s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.493967] env[61991]: INFO nova.compute.manager [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Migrating [ 731.494318] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.494469] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.496068] env[61991]: DEBUG oslo_vmware.api [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129522, 'name': RemoveSnapshot_Task, 'duration_secs': 0.84548} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.496456] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.292s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.498024] env[61991]: INFO nova.compute.claims [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.504020] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 731.504020] env[61991]: INFO nova.compute.manager [None req-a143ad7b-c0cd-4a98-8074-b0a560b6bfaf tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Took 15.53 seconds to snapshot the instance on the hypervisor. [ 731.513690] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129524, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585349} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.513921] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 80d05278-b516-4408-94b0-11bc93500b5c/80d05278-b516-4408-94b0-11bc93500b5c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 731.514178] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.514424] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4422938-f75e-4e26-babd-13aef11de3eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.521184] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 731.521184] env[61991]: value = "task-1129527" [ 731.521184] env[61991]: _type = "Task" [ 731.521184] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.529117] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.609292] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.620182] env[61991]: DEBUG nova.network.neutron [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.768440] env[61991]: DEBUG nova.compute.manager [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Received event network-changed-dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.768644] env[61991]: DEBUG nova.compute.manager [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Refreshing instance network info cache due to event network-changed-dff90314-9bf7-4448-980c-047fc0a617ec. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 731.768863] env[61991]: DEBUG oslo_concurrency.lockutils [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.769015] env[61991]: DEBUG oslo_concurrency.lockutils [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.769622] env[61991]: DEBUG nova.network.neutron [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Refreshing network info cache for port dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 731.814101] env[61991]: DEBUG nova.network.neutron [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Updating instance_info_cache with network_info: [{"id": "8aa50c96-062e-4d6c-9aa0-990a50838237", "address": "fa:16:3e:62:a0:f9", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aa50c96-06", "ovs_interfaceid": "8aa50c96-062e-4d6c-9aa0-990a50838237", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.976668] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 731.996985] env[61991]: INFO nova.compute.rpcapi [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 731.997501] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.035600] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165349} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.035600] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.036951] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9833873-5df0-4de3-8819-1c1f1bf01b74 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.059093] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 80d05278-b516-4408-94b0-11bc93500b5c/80d05278-b516-4408-94b0-11bc93500b5c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.060017] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-421e4f07-3c73-4594-9fd0-d414a67e62e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.080319] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 732.080319] env[61991]: value = "task-1129528" [ 732.080319] env[61991]: _type = "Task" [ 732.080319] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.090484] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129528, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.108031] env[61991]: DEBUG oslo_vmware.api [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Task: {'id': task-1129526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.566967} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.108370] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.108599] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 732.108783] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.108959] env[61991]: INFO nova.compute.manager [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Took 1.73 seconds to destroy the instance on the hypervisor. [ 732.109217] env[61991]: DEBUG oslo.service.loopingcall [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.109406] env[61991]: DEBUG nova.compute.manager [-] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.109500] env[61991]: DEBUG nova.network.neutron [-] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 732.317602] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.319192] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance network_info: |[{"id": "8aa50c96-062e-4d6c-9aa0-990a50838237", "address": "fa:16:3e:62:a0:f9", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aa50c96-06", "ovs_interfaceid": "8aa50c96-062e-4d6c-9aa0-990a50838237", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 732.319787] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:a0:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8aa50c96-062e-4d6c-9aa0-990a50838237', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.328010] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating folder: Project (4a18b2a73f7643e1b89c9660a770e9da). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.328388] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd737c38-d336-46e2-b9c7-710be8839b26 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.340693] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Created folder: Project (4a18b2a73f7643e1b89c9660a770e9da) in parent group-v246753. [ 732.340889] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating folder: Instances. Parent ref: group-v246828. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.341201] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03dd7ca8-7cab-4f90-a536-13d60b54b77c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.350355] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Created folder: Instances in parent group-v246828. [ 732.350635] env[61991]: DEBUG oslo.service.loopingcall [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.350851] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 732.351109] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0017594c-858e-4df9-8b93-5f418aaae906 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.375888] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.375888] env[61991]: value = "task-1129531" [ 732.375888] env[61991]: _type = "Task" [ 732.375888] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.384327] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129531, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.465114] env[61991]: DEBUG nova.compute.manager [req-87de2c37-e6ff-4a10-8803-2285bf78eda8 req-0211d292-8923-421a-b888-11c7fa00b6c1 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Received event network-vif-deleted-227fb866-9358-46ef-a164-201f66f64060 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.465567] env[61991]: INFO nova.compute.manager [req-87de2c37-e6ff-4a10-8803-2285bf78eda8 req-0211d292-8923-421a-b888-11c7fa00b6c1 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Neutron deleted interface 227fb866-9358-46ef-a164-201f66f64060; detaching it from the instance and deleting it from the info cache [ 732.465743] env[61991]: DEBUG nova.network.neutron [req-87de2c37-e6ff-4a10-8803-2285bf78eda8 req-0211d292-8923-421a-b888-11c7fa00b6c1 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.511089] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.519666] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.519899] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.520117] env[61991]: DEBUG nova.network.neutron [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.593588] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129528, 'name': ReconfigVM_Task, 'duration_secs': 0.439513} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.594645] env[61991]: DEBUG nova.network.neutron [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updated VIF entry in instance network info cache for port dff90314-9bf7-4448-980c-047fc0a617ec. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 732.595023] env[61991]: DEBUG nova.network.neutron [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.596402] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 80d05278-b516-4408-94b0-11bc93500b5c/80d05278-b516-4408-94b0-11bc93500b5c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.598015] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b806dff-5320-436f-bb0b-af04b1a4e6a6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.607853] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 732.607853] env[61991]: value = "task-1129532" [ 732.607853] env[61991]: _type = "Task" [ 732.607853] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.623037] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129532, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.874017] env[61991]: DEBUG nova.network.neutron [-] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.886429] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129531, 'name': CreateVM_Task, 'duration_secs': 0.356319} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.888747] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 732.890065] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.890224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.890895] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 732.891025] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-288bb666-8336-41b1-aaa6-e1971611a530 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.895607] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 732.895607] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5272bad8-cc5a-25d7-a0f5-53823827c48e" [ 732.895607] env[61991]: _type = "Task" [ 732.895607] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.900037] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197231ca-9b9e-420f-896f-6d9b4c002efe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.908520] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5272bad8-cc5a-25d7-a0f5-53823827c48e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.910982] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c32054-5635-4900-ae78-d75a3b99a4a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.941987] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499e594f-6884-4b9f-8943-c4020f22f9de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.949705] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54948c5e-7e30-4cd5-9d4c-0d5532e14449 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.963742] env[61991]: DEBUG nova.compute.provider_tree [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.968183] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a183491-0c8a-43fd-8d35-7b586ef8b8ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.977667] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f0f416-79c1-4d51-9c84-92bfb9d8bff2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.004894] env[61991]: DEBUG nova.compute.manager [req-87de2c37-e6ff-4a10-8803-2285bf78eda8 req-0211d292-8923-421a-b888-11c7fa00b6c1 service nova] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Detach interface failed, port_id=227fb866-9358-46ef-a164-201f66f64060, reason: Instance 30bd843a-9d86-4de1-b011-600e2a93bcb3 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 733.097876] env[61991]: DEBUG oslo_concurrency.lockutils [req-b0724b93-6fd3-4ecc-b6e3-38a29603afce req-c7a59bb8-9da3-4f8d-be50-853016229410 service nova] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.120706] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129532, 'name': Rename_Task, 'duration_secs': 0.170438} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.121136] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 733.121254] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a970db9-a4ef-435a-ac1f-667ca2c76978 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.127531] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 733.127531] env[61991]: value = "task-1129533" [ 733.127531] env[61991]: _type = "Task" [ 733.127531] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.136746] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129533, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.248847] env[61991]: DEBUG nova.network.neutron [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.376541] env[61991]: INFO nova.compute.manager [-] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Took 1.27 seconds to deallocate network for instance. [ 733.405853] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5272bad8-cc5a-25d7-a0f5-53823827c48e, 'name': SearchDatastore_Task, 'duration_secs': 0.02152} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.406208] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.406465] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 733.406703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.406850] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.407043] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 733.407311] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d6fb92d-6c3c-469d-a11f-b89e29a1ddf5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.418302] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 733.418302] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 733.420056] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1265f479-e637-4317-8b76-32baf5bbc834 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.424251] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 733.424251] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52053688-cb3c-6a6d-6df8-a2db4eccc6d6" [ 733.424251] env[61991]: _type = "Task" [ 733.424251] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.432416] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52053688-cb3c-6a6d-6df8-a2db4eccc6d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.467096] env[61991]: DEBUG nova.scheduler.client.report [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.637797] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129533, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.753805] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.883714] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.934364] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52053688-cb3c-6a6d-6df8-a2db4eccc6d6, 'name': SearchDatastore_Task, 'duration_secs': 0.045709} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.935227] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37e721d8-6674-4a0f-ad05-ae278c006b80 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.940524] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 733.940524] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286e9ff-d238-db6f-4300-09045786585b" [ 733.940524] env[61991]: _type = "Task" [ 733.940524] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.948114] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286e9ff-d238-db6f-4300-09045786585b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.974030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.974030] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.978922] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.033s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.978922] env[61991]: DEBUG nova.objects.instance [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lazy-loading 'resources' on Instance uuid fe78a8de-6174-47b3-9271-171c4ba76068 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 734.140143] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129533, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.451204] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286e9ff-d238-db6f-4300-09045786585b, 'name': SearchDatastore_Task, 'duration_secs': 0.049349} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.451542] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.451840] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 78631e83-e574-4a90-b062-176b2d12d681/78631e83-e574-4a90-b062-176b2d12d681.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 734.452129] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da200616-ba7b-4abf-8de5-2dd93e3735a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.460120] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 734.460120] env[61991]: value = "task-1129534" [ 734.460120] env[61991]: _type = "Task" [ 734.460120] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.467460] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.483013] env[61991]: DEBUG nova.compute.utils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.484732] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.484848] env[61991]: DEBUG nova.network.neutron [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.532660] env[61991]: DEBUG nova.policy [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a800b157c4984c48b189597954ae3a92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70b65e150f4f4a6d9b20efebe0cb73ac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.647383] env[61991]: DEBUG oslo_vmware.api [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129533, 'name': PowerOnVM_Task, 'duration_secs': 1.058593} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.647812] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 734.648119] env[61991]: INFO nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Took 10.14 seconds to spawn the instance on the hypervisor. [ 734.648475] env[61991]: DEBUG nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 734.649762] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098c3820-e559-43ba-89f6-32e192b7516d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.871899] env[61991]: DEBUG nova.network.neutron [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Successfully created port: 2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.890231] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d9ce77-b335-455b-a385-7f75c94d2039 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.900631] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf46fb54-8682-418b-b667-f845e16ad490 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.934893] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d77b48c-ebf6-4667-a45a-3e3fb4711080 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.940400] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1de0aac-f538-4067-bda4-4561fc2a43a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.954834] env[61991]: DEBUG nova.compute.provider_tree [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.968263] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129534, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462922} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.969170] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 78631e83-e574-4a90-b062-176b2d12d681/78631e83-e574-4a90-b062-176b2d12d681.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 734.969392] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.969636] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57ecac93-d3da-4d35-9999-dc0394c0faa1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.975752] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 734.975752] env[61991]: value = "task-1129535" [ 734.975752] env[61991]: _type = "Task" [ 734.975752] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.983185] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129535, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.987846] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.174232] env[61991]: INFO nova.compute.manager [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Took 32.92 seconds to build instance. [ 735.272577] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671664b4-7d0f-4d0d-a345-740df55f06fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.292211] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 735.461018] env[61991]: DEBUG nova.scheduler.client.report [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.487854] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129535, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06284} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.488210] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.488906] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94f3bb8-a910-4272-8742-0d778331a9b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.517656] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 78631e83-e574-4a90-b062-176b2d12d681/78631e83-e574-4a90-b062-176b2d12d681.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.518065] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3544848-85de-4fc0-8463-2c9f8bf71572 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.542303] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 735.542303] env[61991]: value = "task-1129536" [ 735.542303] env[61991]: _type = "Task" [ 735.542303] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.550598] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129536, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.676175] env[61991]: DEBUG oslo_concurrency.lockutils [None req-135670f1-a087-43f7-a1a4-ce10079b5585 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "80d05278-b516-4408-94b0-11bc93500b5c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.621s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.799757] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 735.799757] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-986a3fe8-a9c1-48cf-9934-6c83f924fed9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.807028] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 735.807028] env[61991]: value = "task-1129537" [ 735.807028] env[61991]: _type = "Task" [ 735.807028] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.816820] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.902450] env[61991]: DEBUG nova.compute.manager [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-changed-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.902677] env[61991]: DEBUG nova.compute.manager [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing instance network info cache due to event network-changed-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.903182] env[61991]: DEBUG oslo_concurrency.lockutils [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.903352] env[61991]: DEBUG oslo_concurrency.lockutils [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.904162] env[61991]: DEBUG nova.network.neutron [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing network info cache for port ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 735.964671] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.988s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.967953] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.796s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.968345] env[61991]: DEBUG nova.objects.instance [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lazy-loading 'resources' on Instance uuid 527561af-ba19-447f-a4ee-618824b9ce0f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 735.993955] env[61991]: INFO nova.scheduler.client.report [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleted allocations for instance fe78a8de-6174-47b3-9271-171c4ba76068 [ 736.000250] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.027592] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.027905] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.028084] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.028284] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.028442] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.028591] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.028801] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.028964] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.029155] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.029325] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.029521] env[61991]: DEBUG nova.virt.hardware [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.030490] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7556a437-41f5-4830-8e61-f7082cb957d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.039277] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fbb331-e237-4b53-95f5-30eb364c70fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.059648] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129536, 'name': ReconfigVM_Task, 'duration_secs': 0.279462} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.060054] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 78631e83-e574-4a90-b062-176b2d12d681/78631e83-e574-4a90-b062-176b2d12d681.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.060781] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29d30db6-efc0-4cc8-9410-98c5e67283c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.067625] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 736.067625] env[61991]: value = "task-1129538" [ 736.067625] env[61991]: _type = "Task" [ 736.067625] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.075911] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129538, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.179295] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.322238] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129537, 'name': PowerOffVM_Task, 'duration_secs': 0.164795} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.322641] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 736.322910] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 736.478783] env[61991]: DEBUG nova.network.neutron [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Successfully updated port: 2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.510081] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41447f63-f02c-4f07-b673-1bfcd5b7e157 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "fe78a8de-6174-47b3-9271-171c4ba76068" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.289s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.578006] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129538, 'name': Rename_Task, 'duration_secs': 0.147322} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.582549] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 736.582997] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b21c7caa-9e09-4b90-83c3-68097563b8e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.589364] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 736.589364] env[61991]: value = "task-1129539" [ 736.589364] env[61991]: _type = "Task" [ 736.589364] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.600526] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129539, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.700349] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.711100] env[61991]: DEBUG nova.network.neutron [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updated VIF entry in instance network info cache for port ea66ec35-3e8b-40d2-baa8-a6d811b6e7be. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 736.711100] env[61991]: DEBUG nova.network.neutron [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.775665] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "acef131a-4263-42ad-964e-bb1e0cb21eec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.776049] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "acef131a-4263-42ad-964e-bb1e0cb21eec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.776348] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "acef131a-4263-42ad-964e-bb1e0cb21eec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.776684] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "acef131a-4263-42ad-964e-bb1e0cb21eec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.776959] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "acef131a-4263-42ad-964e-bb1e0cb21eec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.779461] env[61991]: INFO nova.compute.manager [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Terminating instance [ 736.781811] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "refresh_cache-acef131a-4263-42ad-964e-bb1e0cb21eec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.781989] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquired lock "refresh_cache-acef131a-4263-42ad-964e-bb1e0cb21eec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.782292] env[61991]: DEBUG nova.network.neutron [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.831963] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.832320] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.832521] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.832742] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.833311] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.833311] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.833311] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.833495] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.833687] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.833875] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.834107] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.842553] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-846c3243-f0a6-42b7-98ea-25801b944e49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.859825] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 736.859825] env[61991]: value = "task-1129540" [ 736.859825] env[61991]: _type = "Task" [ 736.859825] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.871762] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129540, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.885781] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67ad85a-328a-4e6e-8172-050d6136181b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.892988] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ccf96a9-f4fc-481f-8fd0-4243aa79cd8a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.923255] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e72228-fe53-46e3-a340-e9ed575bbf5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.931860] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cd19d5-80cd-4b9b-830f-23ce3b261f43 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.947297] env[61991]: DEBUG nova.compute.provider_tree [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.982482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.982654] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquired lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.982740] env[61991]: DEBUG nova.network.neutron [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 737.099518] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129539, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.215890] env[61991]: DEBUG oslo_concurrency.lockutils [req-24c3f282-5a2d-4651-8618-13e4247483c4 req-dc94b01c-9128-4854-8a5f-2fc883321b61 service nova] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.303415] env[61991]: DEBUG nova.network.neutron [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.367559] env[61991]: DEBUG nova.network.neutron [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.372207] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129540, 'name': ReconfigVM_Task, 'duration_secs': 0.160198} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.372768] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 737.453905] env[61991]: DEBUG nova.scheduler.client.report [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.527967] env[61991]: DEBUG nova.network.neutron [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.602304] env[61991]: DEBUG oslo_vmware.api [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129539, 'name': PowerOnVM_Task, 'duration_secs': 0.685572} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.602587] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 737.602786] env[61991]: INFO nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Took 9.50 seconds to spawn the instance on the hypervisor. [ 737.602969] env[61991]: DEBUG nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 737.603761] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205e03ab-41d6-4cbe-9518-392c815d3274 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.684918] env[61991]: DEBUG nova.network.neutron [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Updating instance_info_cache with network_info: [{"id": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "address": "fa:16:3e:dd:75:80", "network": {"id": "32f62b4b-b851-4224-930f-b29370cd8309", "bridge": "br-int", "label": "tempest-ServersTestJSON-231629090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70b65e150f4f4a6d9b20efebe0cb73ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e472308-0d", "ovs_interfaceid": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.873640] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Releasing lock "refresh_cache-acef131a-4263-42ad-964e-bb1e0cb21eec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.874235] env[61991]: DEBUG nova.compute.manager [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.875028] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.876059] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004c75e0-268e-44ba-b9a0-f5a07cc558ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.881687] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:18:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b41f5b31-f99f-43b3-b2c2-a0cd889c0027',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-424611864',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.881971] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.882225] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.882475] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.882671] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.882859] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.883131] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.883351] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.883572] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.883782] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.883994] env[61991]: DEBUG nova.virt.hardware [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.889911] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 737.890224] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca8c0c8c-641b-4ad8-80f9-24118fad9399 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.910106] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 737.911434] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54f32d1b-8f80-4119-b5ba-f2bc94e23b27 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.913129] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 737.913129] env[61991]: value = "task-1129541" [ 737.913129] env[61991]: _type = "Task" [ 737.913129] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.917588] env[61991]: DEBUG oslo_vmware.api [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 737.917588] env[61991]: value = "task-1129542" [ 737.917588] env[61991]: _type = "Task" [ 737.917588] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.924622] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.928948] env[61991]: DEBUG nova.compute.manager [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Received event network-vif-plugged-2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.928948] env[61991]: DEBUG oslo_concurrency.lockutils [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] Acquiring lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.929056] env[61991]: DEBUG oslo_concurrency.lockutils [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.929404] env[61991]: DEBUG oslo_concurrency.lockutils [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.929625] env[61991]: DEBUG nova.compute.manager [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] No waiting events found dispatching network-vif-plugged-2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 737.929845] env[61991]: WARNING nova.compute.manager [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Received unexpected event network-vif-plugged-2e472308-0d44-4e9e-ae6e-3afaad4796b9 for instance with vm_state building and task_state spawning. [ 737.930078] env[61991]: DEBUG nova.compute.manager [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Received event network-changed-2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.930258] env[61991]: DEBUG nova.compute.manager [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Refreshing instance network info cache due to event network-changed-2e472308-0d44-4e9e-ae6e-3afaad4796b9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.930472] env[61991]: DEBUG oslo_concurrency.lockutils [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] Acquiring lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.934413] env[61991]: DEBUG oslo_vmware.api [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.961713] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.962984] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.509s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.965170] env[61991]: INFO nova.compute.claims [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.982099] env[61991]: INFO nova.scheduler.client.report [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Deleted allocations for instance 527561af-ba19-447f-a4ee-618824b9ce0f [ 738.121512] env[61991]: INFO nova.compute.manager [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Took 33.88 seconds to build instance. [ 738.189033] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Releasing lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.189033] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Instance network_info: |[{"id": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "address": "fa:16:3e:dd:75:80", "network": {"id": "32f62b4b-b851-4224-930f-b29370cd8309", "bridge": "br-int", "label": "tempest-ServersTestJSON-231629090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70b65e150f4f4a6d9b20efebe0cb73ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e472308-0d", "ovs_interfaceid": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 738.189033] env[61991]: DEBUG oslo_concurrency.lockutils [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] Acquired lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.189277] env[61991]: DEBUG nova.network.neutron [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Refreshing network info cache for port 2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.190403] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:75:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea00b53a-9c9b-4592-ab95-7e10473f338d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e472308-0d44-4e9e-ae6e-3afaad4796b9', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 738.198587] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Creating folder: Project (70b65e150f4f4a6d9b20efebe0cb73ac). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 738.199569] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0149dd1-1b79-4cfe-b908-87a6974140f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.210998] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Created folder: Project (70b65e150f4f4a6d9b20efebe0cb73ac) in parent group-v246753. [ 738.211215] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Creating folder: Instances. Parent ref: group-v246831. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 738.211442] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af73faa6-14ce-4c60-9bc8-5817b28f6915 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.219877] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Created folder: Instances in parent group-v246831. [ 738.220064] env[61991]: DEBUG oslo.service.loopingcall [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.220252] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 738.220450] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ff49b22-eea5-4b09-b9fb-ec30ffac3880 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.238635] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 738.238635] env[61991]: value = "task-1129545" [ 738.238635] env[61991]: _type = "Task" [ 738.238635] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.246095] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129545, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.424990] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129541, 'name': ReconfigVM_Task, 'duration_secs': 0.219727} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.425658] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 738.426447] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35dfb85-25b8-483d-a966-060531ad82ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.431473] env[61991]: DEBUG oslo_vmware.api [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129542, 'name': PowerOffVM_Task, 'duration_secs': 0.179252} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.432016] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 738.432230] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 738.432477] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebc1659a-2564-4447-8d47-d82dc6c095c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.452272] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.452880] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb5045b5-63a5-41f7-adbf-f9cfe78c9dbe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.473986] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 738.473986] env[61991]: value = "task-1129547" [ 738.473986] env[61991]: _type = "Task" [ 738.473986] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.475229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 738.475459] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 738.475645] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleting the datastore file [datastore2] acef131a-4263-42ad-964e-bb1e0cb21eec {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 738.479007] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7e5ee74-386f-4a36-ae7e-f1ff49594272 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.487373] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129547, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.491283] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9f4287a7-932e-4b70-a13d-da1c3ecbaf10 tempest-InstanceActionsNegativeTestJSON-1335680081 tempest-InstanceActionsNegativeTestJSON-1335680081-project-member] Lock "527561af-ba19-447f-a4ee-618824b9ce0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.317s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.492763] env[61991]: DEBUG oslo_vmware.api [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for the task: (returnval){ [ 738.492763] env[61991]: value = "task-1129548" [ 738.492763] env[61991]: _type = "Task" [ 738.492763] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.501016] env[61991]: DEBUG oslo_vmware.api [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129548, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.624183] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ca4fdd2-672e-465d-afc5-f1b562e0c753 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.141s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.750881] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129545, 'name': CreateVM_Task, 'duration_secs': 0.465397} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.751070] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 738.751761] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.751992] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.752289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 738.752564] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e49c86e7-eb37-4df7-9f18-ff46a82d32ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.756961] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 738.756961] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527b4f59-5b78-6542-f205-7e4aff5deef9" [ 738.756961] env[61991]: _type = "Task" [ 738.756961] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.764461] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527b4f59-5b78-6542-f205-7e4aff5deef9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.987095] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129547, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.001926] env[61991]: DEBUG oslo_vmware.api [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Task: {'id': task-1129548, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121757} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.002324] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.002691] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 739.002925] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.003177] env[61991]: INFO nova.compute.manager [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Took 1.13 seconds to destroy the instance on the hypervisor. [ 739.003517] env[61991]: DEBUG oslo.service.loopingcall [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.003797] env[61991]: DEBUG nova.compute.manager [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.004012] env[61991]: DEBUG nova.network.neutron [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.024819] env[61991]: DEBUG nova.network.neutron [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.128659] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.269608] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527b4f59-5b78-6542-f205-7e4aff5deef9, 'name': SearchDatastore_Task, 'duration_secs': 0.010969} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.270901] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.270901] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.270901] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.270901] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.270901] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.271136] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bf398bb-e9af-43f1-937b-3438063b6e52 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.280906] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.281246] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 739.282165] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cafee9c0-31c3-4c2c-8a4c-00a76ef4100e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.287585] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 739.287585] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52716e54-74d4-89c0-80f3-e98fa1e6ab0f" [ 739.287585] env[61991]: _type = "Task" [ 739.287585] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.297460] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52716e54-74d4-89c0-80f3-e98fa1e6ab0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.317155] env[61991]: DEBUG nova.network.neutron [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Updated VIF entry in instance network info cache for port 2e472308-0d44-4e9e-ae6e-3afaad4796b9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 739.317465] env[61991]: DEBUG nova.network.neutron [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Updating instance_info_cache with network_info: [{"id": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "address": "fa:16:3e:dd:75:80", "network": {"id": "32f62b4b-b851-4224-930f-b29370cd8309", "bridge": "br-int", "label": "tempest-ServersTestJSON-231629090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70b65e150f4f4a6d9b20efebe0cb73ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e472308-0d", "ovs_interfaceid": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.364349] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e7c128-bb1e-4340-aff0-d2585b5d405b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.372762] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6a84a3-e3fb-4266-b0f2-617edbac4652 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.408203] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7aea68-efa7-42b1-b8b6-69dd80a429d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.416660] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41478fcb-7593-4b9e-92b0-2b1d187a3e99 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.432033] env[61991]: DEBUG nova.compute.provider_tree [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.486915] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129547, 'name': ReconfigVM_Task, 'duration_secs': 0.519979} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.487367] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfigured VM instance instance-0000002b to attach disk [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.487648] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 739.528593] env[61991]: DEBUG nova.network.neutron [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.657020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.800890] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52716e54-74d4-89c0-80f3-e98fa1e6ab0f, 'name': SearchDatastore_Task, 'duration_secs': 0.038171} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.801767] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8798289-e4c7-4a4f-a2ba-8544256bea48 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.807509] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 739.807509] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d08873-2514-8744-c1b9-4f1203bf5442" [ 739.807509] env[61991]: _type = "Task" [ 739.807509] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.817029] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d08873-2514-8744-c1b9-4f1203bf5442, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.820787] env[61991]: DEBUG oslo_concurrency.lockutils [req-222a9f11-1518-4d9c-b00d-455b5f8ef7b2 req-33d5af25-f99f-4756-9bce-ac7a19ed72a9 service nova] Releasing lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.938689] env[61991]: DEBUG nova.scheduler.client.report [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.995127] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02b7c93-7d8b-4db7-b5fb-c765a547c247 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.019875] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6f0eb2-130b-48db-8158-f3eb3d29ee98 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.041513] env[61991]: INFO nova.compute.manager [-] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Took 1.04 seconds to deallocate network for instance. [ 740.045021] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 740.170020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "78631e83-e574-4a90-b062-176b2d12d681" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.170020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.170020] env[61991]: INFO nova.compute.manager [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Shelving [ 740.321723] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d08873-2514-8744-c1b9-4f1203bf5442, 'name': SearchDatastore_Task, 'duration_secs': 0.010861} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.322076] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.322352] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 700ce4a7-d597-449a-9379-0cfb0c8f82fc/700ce4a7-d597-449a-9379-0cfb0c8f82fc.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 740.322619] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4768123-34e2-4d2f-b866-3e4e09289cf6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.329788] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 740.329788] env[61991]: value = "task-1129549" [ 740.329788] env[61991]: _type = "Task" [ 740.329788] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.337987] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.447031] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.447031] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.450782] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 26.890s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.450953] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.451115] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 740.451571] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.511s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.453989] env[61991]: INFO nova.compute.claims [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.458621] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d915064-977c-48df-95e6-f386dbf044d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.467377] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb151c6-de3f-404b-a50e-4b61f08202e1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.482228] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835dcf0d-6b56-48aa-be0e-49b3492b7207 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.490018] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac9029d-26b0-4e45-9cc4-dacbad6e7590 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.520113] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181404MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 740.520281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.557226] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.597579] env[61991]: DEBUG nova.network.neutron [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Port 863ee2c0-0c5a-49c8-a443-3c1c3d73c458 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 740.686954] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 740.687358] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a275bd7-b727-4748-afcb-780f47b006bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.694481] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 740.694481] env[61991]: value = "task-1129550" [ 740.694481] env[61991]: _type = "Task" [ 740.694481] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.704947] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129550, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.844745] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129549, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.959774] env[61991]: DEBUG nova.compute.utils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.961316] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.962313] env[61991]: DEBUG nova.network.neutron [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 741.068953] env[61991]: DEBUG nova.policy [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b10ac4cad934e7ea793120929ad84ae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ddab761ff5f49aeb8a4b611cd9d7603', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.203385] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129550, 'name': PowerOffVM_Task, 'duration_secs': 0.201542} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.203670] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 741.204522] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336ad391-8094-437e-b4d2-1e1b1ee2a77d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.223902] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40188c6-9072-482b-bb65-3c8b8702c14e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.343243] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.831398} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.343533] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 700ce4a7-d597-449a-9379-0cfb0c8f82fc/700ce4a7-d597-449a-9379-0cfb0c8f82fc.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.343885] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.344163] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd7dd1d0-6d76-429e-9b5e-006273d7ce32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.351197] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 741.351197] env[61991]: value = "task-1129551" [ 741.351197] env[61991]: _type = "Task" [ 741.351197] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.359684] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129551, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.469668] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.480981] env[61991]: DEBUG nova.network.neutron [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Successfully created port: 29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.632989] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.633395] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.633593] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.734163] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 741.734533] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e33a4aae-9cf9-4f42-a145-f38c0fda09d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.742684] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 741.742684] env[61991]: value = "task-1129552" [ 741.742684] env[61991]: _type = "Task" [ 741.742684] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.750402] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129552, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.861156] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129551, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.141568} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.863166] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 741.864618] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39ada6c-aeec-4087-a3ce-6dd47841d4f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.892249] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 700ce4a7-d597-449a-9379-0cfb0c8f82fc/700ce4a7-d597-449a-9379-0cfb0c8f82fc.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 741.893603] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e4a59a2-0cab-4039-9f3b-22c6e23a4c59 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.908901] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5840fd3-03cf-456f-9860-ad5d1340be22 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.917656] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac5b150-7666-4b73-b18a-c52445ed4d13 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.920927] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 741.920927] env[61991]: value = "task-1129553" [ 741.920927] env[61991]: _type = "Task" [ 741.920927] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.951256] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2467db7b-cd08-4b1e-bd86-f8c618031e9a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.956993] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129553, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.962020] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23f82c1-9bfb-43e2-9f11-1d49ec2f766e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.976460] env[61991]: DEBUG nova.compute.provider_tree [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.252476] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129552, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.434162] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129553, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.478641] env[61991]: DEBUG nova.scheduler.client.report [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.482862] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.517976] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.518258] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.518421] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.518605] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.518752] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.518900] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.519117] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.519281] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.519448] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.519611] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.519782] env[61991]: DEBUG nova.virt.hardware [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.520648] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd14243-f737-4316-9730-4acac661891d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.530509] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19c2371-bdd5-4ec5-ac29-72beec669e00 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.726503] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.726503] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.726621] env[61991]: DEBUG nova.network.neutron [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.755045] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129552, 'name': CreateSnapshot_Task, 'duration_secs': 0.974708} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.755045] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 742.755704] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebc864e-1e6b-479a-a689-d230a7ab2e3b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.933742] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129553, 'name': ReconfigVM_Task, 'duration_secs': 0.67908} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.934045] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 700ce4a7-d597-449a-9379-0cfb0c8f82fc/700ce4a7-d597-449a-9379-0cfb0c8f82fc.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.934728] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9732f96e-44c4-4bb5-b1be-2029416b6c3c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.942269] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 742.942269] env[61991]: value = "task-1129554" [ 742.942269] env[61991]: _type = "Task" [ 742.942269] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.952756] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129554, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.968398] env[61991]: DEBUG nova.compute.manager [req-bd6ff757-ca3a-4ee8-a25c-fb78c0be8293 req-8d6e074a-f550-46c3-98fb-cb7a888cfd4f service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Received event network-vif-plugged-29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.968679] env[61991]: DEBUG oslo_concurrency.lockutils [req-bd6ff757-ca3a-4ee8-a25c-fb78c0be8293 req-8d6e074a-f550-46c3-98fb-cb7a888cfd4f service nova] Acquiring lock "3e06e424-b64e-4fd2-8013-27760200c41d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.968927] env[61991]: DEBUG oslo_concurrency.lockutils [req-bd6ff757-ca3a-4ee8-a25c-fb78c0be8293 req-8d6e074a-f550-46c3-98fb-cb7a888cfd4f service nova] Lock "3e06e424-b64e-4fd2-8013-27760200c41d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.969299] env[61991]: DEBUG oslo_concurrency.lockutils [req-bd6ff757-ca3a-4ee8-a25c-fb78c0be8293 req-8d6e074a-f550-46c3-98fb-cb7a888cfd4f service nova] Lock "3e06e424-b64e-4fd2-8013-27760200c41d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.969559] env[61991]: DEBUG nova.compute.manager [req-bd6ff757-ca3a-4ee8-a25c-fb78c0be8293 req-8d6e074a-f550-46c3-98fb-cb7a888cfd4f service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] No waiting events found dispatching network-vif-plugged-29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 742.969779] env[61991]: WARNING nova.compute.manager [req-bd6ff757-ca3a-4ee8-a25c-fb78c0be8293 req-8d6e074a-f550-46c3-98fb-cb7a888cfd4f service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Received unexpected event network-vif-plugged-29738db6-7735-48b9-920a-e9d98d76aa24 for instance with vm_state building and task_state spawning. [ 742.986668] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.987241] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 742.990417] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.810s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.990690] env[61991]: DEBUG nova.objects.instance [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lazy-loading 'resources' on Instance uuid 43cb61e7-1748-40d9-a287-1179c8219c2a {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 743.075301] env[61991]: DEBUG nova.network.neutron [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Successfully updated port: 29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 743.279085] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 743.279526] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ae8e0c24-6c8c-49e7-a087-03ebf3581282 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.288042] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 743.288042] env[61991]: value = "task-1129555" [ 743.288042] env[61991]: _type = "Task" [ 743.288042] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.295987] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129555, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.453241] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129554, 'name': Rename_Task, 'duration_secs': 0.132984} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.453514] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 743.453764] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d5191f5-75df-4a89-83e2-0817a0669c26 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.461582] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 743.461582] env[61991]: value = "task-1129556" [ 743.461582] env[61991]: _type = "Task" [ 743.461582] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.468102] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129556, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.495602] env[61991]: DEBUG nova.compute.utils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.501331] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 743.501331] env[61991]: DEBUG nova.network.neutron [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 743.535366] env[61991]: DEBUG nova.network.neutron [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.580106] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.580434] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.580761] env[61991]: DEBUG nova.network.neutron [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.598739] env[61991]: DEBUG nova.policy [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbb68bcfad394700b365bbd5bb60b098', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53dfdbbaa1094205b865d3fcf591136e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 743.801497] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129555, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.885266] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68664d66-9583-4941-8e6a-cac265545592 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.893381] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d037fb-c8e3-4a3f-9119-c4d9136e003f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.929181] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332172aa-6f0a-4c1b-ae20-cd6001465fe1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.937538] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc209e55-4ad7-4edb-bc2a-285100df7a96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.952099] env[61991]: DEBUG nova.compute.provider_tree [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.970932] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129556, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.001195] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.045462] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.128686] env[61991]: DEBUG nova.network.neutron [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.137382] env[61991]: DEBUG nova.network.neutron [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Successfully created port: b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.289039] env[61991]: DEBUG nova.network.neutron [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updating instance_info_cache with network_info: [{"id": "29738db6-7735-48b9-920a-e9d98d76aa24", "address": "fa:16:3e:67:8e:c4", "network": {"id": "7bdca730-2d82-430c-9297-63bf61dec70e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1406357972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ddab761ff5f49aeb8a4b611cd9d7603", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29738db6-77", "ovs_interfaceid": "29738db6-7735-48b9-920a-e9d98d76aa24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.299159] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129555, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.456974] env[61991]: DEBUG nova.scheduler.client.report [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.477492] env[61991]: DEBUG oslo_vmware.api [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129556, 'name': PowerOnVM_Task, 'duration_secs': 0.764236} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.477882] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 744.477997] env[61991]: INFO nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Took 8.48 seconds to spawn the instance on the hypervisor. [ 744.478170] env[61991]: DEBUG nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 744.478988] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330146d4-4047-4f62-858a-d4a526df8a56 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.572066] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e8f25e-376c-4b0d-bf26-2ecda54f496f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.595922] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a0e707-8f2c-4664-86d4-3a84bc80825e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.604657] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 744.791912] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Releasing lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.792581] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Instance network_info: |[{"id": "29738db6-7735-48b9-920a-e9d98d76aa24", "address": "fa:16:3e:67:8e:c4", "network": {"id": "7bdca730-2d82-430c-9297-63bf61dec70e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1406357972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ddab761ff5f49aeb8a4b611cd9d7603", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29738db6-77", "ovs_interfaceid": "29738db6-7735-48b9-920a-e9d98d76aa24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 744.795969] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:8e:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea4fe416-47a6-4542-b59d-8c71ab4d6503', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29738db6-7735-48b9-920a-e9d98d76aa24', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.803974] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Creating folder: Project (1ddab761ff5f49aeb8a4b611cd9d7603). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 744.808064] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1352d20d-ab54-4344-a538-79ca4b458836 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.813971] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129555, 'name': CloneVM_Task, 'duration_secs': 1.389222} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.815395] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Created linked-clone VM from snapshot [ 744.815781] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Created folder: Project (1ddab761ff5f49aeb8a4b611cd9d7603) in parent group-v246753. [ 744.816038] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Creating folder: Instances. Parent ref: group-v246836. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 744.817059] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3a7da1-fdb8-4a46-9ac5-da1e4c679d7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.820562] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f557a700-4a84-4f39-a7a9-4ddd886156d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.829024] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Uploading image e8cac67d-5662-46ad-be70-bc504bd13ab9 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 744.829934] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Created folder: Instances in parent group-v246836. [ 744.830231] env[61991]: DEBUG oslo.service.loopingcall [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.830434] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 744.830644] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f87563df-3726-4eae-a0bd-65cf5bf93acd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.851477] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.851477] env[61991]: value = "task-1129559" [ 744.851477] env[61991]: _type = "Task" [ 744.851477] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.853460] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 744.853460] env[61991]: value = "vm-246835" [ 744.853460] env[61991]: _type = "VirtualMachine" [ 744.853460] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 744.853783] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-574fd878-46d5-42fa-984f-47d29f306a3b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.868035] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129559, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.870220] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lease: (returnval){ [ 744.870220] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524a7731-d2f8-448f-3c1d-9b10c23026d2" [ 744.870220] env[61991]: _type = "HttpNfcLease" [ 744.870220] env[61991]: } obtained for exporting VM: (result){ [ 744.870220] env[61991]: value = "vm-246835" [ 744.870220] env[61991]: _type = "VirtualMachine" [ 744.870220] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 744.870566] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the lease: (returnval){ [ 744.870566] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524a7731-d2f8-448f-3c1d-9b10c23026d2" [ 744.870566] env[61991]: _type = "HttpNfcLease" [ 744.870566] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 744.877383] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 744.877383] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524a7731-d2f8-448f-3c1d-9b10c23026d2" [ 744.877383] env[61991]: _type = "HttpNfcLease" [ 744.877383] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 744.963672] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.966096] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.296s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.968077] env[61991]: INFO nova.compute.claims [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.005282] env[61991]: INFO nova.compute.manager [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Took 37.82 seconds to build instance. [ 745.008154] env[61991]: INFO nova.scheduler.client.report [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Deleted allocations for instance 43cb61e7-1748-40d9-a287-1179c8219c2a [ 745.013379] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.047069] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.047277] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.047455] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.049063] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.049063] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.049063] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.049063] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.049288] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.049441] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.049683] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.049924] env[61991]: DEBUG nova.virt.hardware [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.052103] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47e331c-2fef-4610-aeae-deb8c94dad0e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.061994] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6521f4b6-b539-4ddf-ac56-fa4bbcf8cd83 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.111763] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 745.112428] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12fc985f-9e67-4b14-b97b-d36a6909ff79 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.120275] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 745.120275] env[61991]: value = "task-1129561" [ 745.120275] env[61991]: _type = "Task" [ 745.120275] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.129704] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.364785] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129559, 'name': CreateVM_Task, 'duration_secs': 0.325755} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.364957] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 745.365756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.365939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.366919] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 745.367567] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca309ea6-977c-43c6-b376-d5dcd654fa28 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.370045] env[61991]: DEBUG nova.compute.manager [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Received event network-changed-29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.370236] env[61991]: DEBUG nova.compute.manager [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Refreshing instance network info cache due to event network-changed-29738db6-7735-48b9-920a-e9d98d76aa24. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 745.370438] env[61991]: DEBUG oslo_concurrency.lockutils [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] Acquiring lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.370601] env[61991]: DEBUG oslo_concurrency.lockutils [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] Acquired lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.370762] env[61991]: DEBUG nova.network.neutron [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Refreshing network info cache for port 29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 745.377452] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 745.377452] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df670a-9564-c978-2fb3-ed99069097b7" [ 745.377452] env[61991]: _type = "Task" [ 745.377452] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.382864] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 745.382864] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524a7731-d2f8-448f-3c1d-9b10c23026d2" [ 745.382864] env[61991]: _type = "HttpNfcLease" [ 745.382864] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 745.383444] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 745.383444] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524a7731-d2f8-448f-3c1d-9b10c23026d2" [ 745.383444] env[61991]: _type = "HttpNfcLease" [ 745.383444] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 745.384146] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da01d7b8-0382-491d-9aef-3be970e2a6ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.391708] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df670a-9564-c978-2fb3-ed99069097b7, 'name': SearchDatastore_Task, 'duration_secs': 0.010846} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.392299] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.392586] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.392862] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.393029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.393219] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.393555] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a10b6ab-92be-4809-80e7-48c5b6130803 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.398214] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286f284-1aa3-ed90-8861-476295d8a854/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 745.398393] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286f284-1aa3-ed90-8861-476295d8a854/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 745.460429] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.460429] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 745.465350] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61c10e20-e2c3-4a1e-a085-390c76dd9745 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.468633] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 745.468633] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df06e7-d4ff-8380-7e6e-37a6efe2264e" [ 745.468633] env[61991]: _type = "Task" [ 745.468633] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.479381] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df06e7-d4ff-8380-7e6e-37a6efe2264e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.507947] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2e4c8f4c-3c60-48c5-ac96-522b68c524f6 tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.186s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.519264] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ce795316-cc03-47fd-ac5b-ab68419f20fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.522558] env[61991]: DEBUG oslo_concurrency.lockutils [None req-37d84f37-0e18-4a98-b6d4-3c9f7556fa96 tempest-ServerPasswordTestJSON-973856191 tempest-ServerPasswordTestJSON-973856191-project-member] Lock "43cb61e7-1748-40d9-a287-1179c8219c2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.281s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.569203] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "884e465b-e14f-4114-81a6-40e2e8694558" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.569376] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "884e465b-e14f-4114-81a6-40e2e8694558" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.631231] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129561, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.986415] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df06e7-d4ff-8380-7e6e-37a6efe2264e, 'name': SearchDatastore_Task, 'duration_secs': 0.009641} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.990785] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-534587de-b7a9-43d1-9ec4-db20ee0b6930 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.998746] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 745.998746] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b5f661-6f54-8761-7269-3bd0f7c2ceb5" [ 745.998746] env[61991]: _type = "Task" [ 745.998746] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.008154] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b5f661-6f54-8761-7269-3bd0f7c2ceb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.011966] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.055085] env[61991]: DEBUG nova.compute.manager [req-f18e19c8-8c05-4229-9714-2a0e2e40612d req-8f58ba44-af12-4af3-8c5f-fe803c523968 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Received event network-vif-plugged-b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.059012] env[61991]: DEBUG oslo_concurrency.lockutils [req-f18e19c8-8c05-4229-9714-2a0e2e40612d req-8f58ba44-af12-4af3-8c5f-fe803c523968 service nova] Acquiring lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.059307] env[61991]: DEBUG oslo_concurrency.lockutils [req-f18e19c8-8c05-4229-9714-2a0e2e40612d req-8f58ba44-af12-4af3-8c5f-fe803c523968 service nova] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.059438] env[61991]: DEBUG oslo_concurrency.lockutils [req-f18e19c8-8c05-4229-9714-2a0e2e40612d req-8f58ba44-af12-4af3-8c5f-fe803c523968 service nova] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.060302] env[61991]: DEBUG nova.compute.manager [req-f18e19c8-8c05-4229-9714-2a0e2e40612d req-8f58ba44-af12-4af3-8c5f-fe803c523968 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] No waiting events found dispatching network-vif-plugged-b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 746.060302] env[61991]: WARNING nova.compute.manager [req-f18e19c8-8c05-4229-9714-2a0e2e40612d req-8f58ba44-af12-4af3-8c5f-fe803c523968 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Received unexpected event network-vif-plugged-b1cd4427-17b1-40c2-8937-03d1f29e3d21 for instance with vm_state building and task_state spawning. [ 746.139910] env[61991]: DEBUG oslo_vmware.api [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129561, 'name': PowerOnVM_Task, 'duration_secs': 0.611212} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.141395] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 746.141860] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0411dd-771f-4c6b-9e05-56808b115b26 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance 'c14d99a3-950d-44d0-b330-3031a1f8a2be' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 746.256669] env[61991]: DEBUG nova.network.neutron [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Successfully updated port: b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 746.406939] env[61991]: DEBUG nova.network.neutron [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updated VIF entry in instance network info cache for port 29738db6-7735-48b9-920a-e9d98d76aa24. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 746.407615] env[61991]: DEBUG nova.network.neutron [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updating instance_info_cache with network_info: [{"id": "29738db6-7735-48b9-920a-e9d98d76aa24", "address": "fa:16:3e:67:8e:c4", "network": {"id": "7bdca730-2d82-430c-9297-63bf61dec70e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1406357972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ddab761ff5f49aeb8a4b611cd9d7603", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29738db6-77", "ovs_interfaceid": "29738db6-7735-48b9-920a-e9d98d76aa24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.453080] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f619a9b-3c92-4ffe-b884-0937a873a38b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.462745] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539c8d62-eeff-4ade-984f-7eccbd749596 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.514363] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d80471-7fef-49c7-9f7c-1630f96f2ecf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.527944] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b5f661-6f54-8761-7269-3bd0f7c2ceb5, 'name': SearchDatastore_Task, 'duration_secs': 0.01329} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.530703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.530987] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 3e06e424-b64e-4fd2-8013-27760200c41d/3e06e424-b64e-4fd2-8013-27760200c41d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 746.532418] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00536b98-8475-4400-9524-44d274f6ff5d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.537822] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8d7e772-d4d6-4518-98e8-b99be274d879 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.555144] env[61991]: DEBUG nova.compute.provider_tree [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.558034] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.560577] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 746.560577] env[61991]: value = "task-1129562" [ 746.560577] env[61991]: _type = "Task" [ 746.560577] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.574574] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.759149] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "refresh_cache-4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.761264] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquired lock "refresh_cache-4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.761264] env[61991]: DEBUG nova.network.neutron [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 746.910304] env[61991]: DEBUG oslo_concurrency.lockutils [req-d3013b67-9a1a-401c-bda2-cac0338eff6c req-0864f9a5-7106-4b99-a9f6-470adcc9e53d service nova] Releasing lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.062438] env[61991]: DEBUG nova.scheduler.client.report [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.081393] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129562, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494761} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.081805] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 3e06e424-b64e-4fd2-8013-27760200c41d/3e06e424-b64e-4fd2-8013-27760200c41d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 747.082100] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.082435] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e779d64e-7a6a-4846-b6d1-c298a00896bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.091148] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 747.091148] env[61991]: value = "task-1129563" [ 747.091148] env[61991]: _type = "Task" [ 747.091148] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.101361] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.300565] env[61991]: DEBUG nova.network.neutron [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.420572] env[61991]: DEBUG nova.compute.manager [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Received event network-changed-2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.420865] env[61991]: DEBUG nova.compute.manager [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Refreshing instance network info cache due to event network-changed-2e472308-0d44-4e9e-ae6e-3afaad4796b9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 747.422067] env[61991]: DEBUG oslo_concurrency.lockutils [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] Acquiring lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.422283] env[61991]: DEBUG oslo_concurrency.lockutils [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] Acquired lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.422841] env[61991]: DEBUG nova.network.neutron [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Refreshing network info cache for port 2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 747.483144] env[61991]: DEBUG nova.network.neutron [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Updating instance_info_cache with network_info: [{"id": "b1cd4427-17b1-40c2-8937-03d1f29e3d21", "address": "fa:16:3e:17:5b:5b", "network": {"id": "334b8767-1629-42bc-94cc-70a52b88b1dc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1266141268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53dfdbbaa1094205b865d3fcf591136e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1cd4427-17", "ovs_interfaceid": "b1cd4427-17b1-40c2-8937-03d1f29e3d21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.575688] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.576251] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.579276] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.833s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.580986] env[61991]: INFO nova.compute.claims [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.602210] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066565} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.603691] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 747.603691] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e1ceb9-9861-437f-b0cb-682fdda9ad85 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.627458] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 3e06e424-b64e-4fd2-8013-27760200c41d/3e06e424-b64e-4fd2-8013-27760200c41d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 747.631216] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3ba5505-bb48-4e85-9f09-b8bd9d46a6fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.654775] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 747.654775] env[61991]: value = "task-1129564" [ 747.654775] env[61991]: _type = "Task" [ 747.654775] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.676602] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129564, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.991282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Releasing lock "refresh_cache-4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.991282] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Instance network_info: |[{"id": "b1cd4427-17b1-40c2-8937-03d1f29e3d21", "address": "fa:16:3e:17:5b:5b", "network": {"id": "334b8767-1629-42bc-94cc-70a52b88b1dc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1266141268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53dfdbbaa1094205b865d3fcf591136e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1cd4427-17", "ovs_interfaceid": "b1cd4427-17b1-40c2-8937-03d1f29e3d21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 747.991282] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:5b:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1cd4427-17b1-40c2-8937-03d1f29e3d21', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 748.002829] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Creating folder: Project (53dfdbbaa1094205b865d3fcf591136e). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 748.003650] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a34e628c-ced8-418b-ac30-078ff0b6c197 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.015421] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Created folder: Project (53dfdbbaa1094205b865d3fcf591136e) in parent group-v246753. [ 748.015774] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Creating folder: Instances. Parent ref: group-v246839. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 748.017390] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2b82e25-8910-4c17-8fb8-043c6c16a3cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.027517] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Created folder: Instances in parent group-v246839. [ 748.027716] env[61991]: DEBUG oslo.service.loopingcall [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.030808] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 748.031105] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0822e3c-df1f-4297-9538-66da273e9f22 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.055102] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 748.055102] env[61991]: value = "task-1129567" [ 748.055102] env[61991]: _type = "Task" [ 748.055102] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.064709] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129567, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.077421] env[61991]: DEBUG nova.compute.manager [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Received event network-changed-b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.077647] env[61991]: DEBUG nova.compute.manager [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Refreshing instance network info cache due to event network-changed-b1cd4427-17b1-40c2-8937-03d1f29e3d21. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.077993] env[61991]: DEBUG oslo_concurrency.lockutils [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] Acquiring lock "refresh_cache-4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.078404] env[61991]: DEBUG oslo_concurrency.lockutils [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] Acquired lock "refresh_cache-4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.078404] env[61991]: DEBUG nova.network.neutron [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Refreshing network info cache for port b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 748.085779] env[61991]: DEBUG nova.compute.utils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.091442] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.091631] env[61991]: DEBUG nova.network.neutron [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 748.142174] env[61991]: DEBUG nova.policy [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64adcb769f9d479ea70a2a71f7c14398', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a0346ea1b8f4cf9a9322dea401e7b43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.164848] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129564, 'name': ReconfigVM_Task, 'duration_secs': 0.48053} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.165219] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 3e06e424-b64e-4fd2-8013-27760200c41d/3e06e424-b64e-4fd2-8013-27760200c41d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.165861] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ce85423-9936-4050-8a49-238ce5d9fefe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.173791] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 748.173791] env[61991]: value = "task-1129568" [ 748.173791] env[61991]: _type = "Task" [ 748.173791] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.184240] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129568, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.293759] env[61991]: DEBUG nova.network.neutron [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Updated VIF entry in instance network info cache for port 2e472308-0d44-4e9e-ae6e-3afaad4796b9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 748.294233] env[61991]: DEBUG nova.network.neutron [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Updating instance_info_cache with network_info: [{"id": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "address": "fa:16:3e:dd:75:80", "network": {"id": "32f62b4b-b851-4224-930f-b29370cd8309", "bridge": "br-int", "label": "tempest-ServersTestJSON-231629090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.136", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70b65e150f4f4a6d9b20efebe0cb73ac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea00b53a-9c9b-4592-ab95-7e10473f338d", "external-id": "nsx-vlan-transportzone-235", "segmentation_id": 235, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e472308-0d", "ovs_interfaceid": "2e472308-0d44-4e9e-ae6e-3afaad4796b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.538046] env[61991]: DEBUG nova.network.neutron [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Successfully created port: 248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.540826] env[61991]: DEBUG nova.network.neutron [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Port 863ee2c0-0c5a-49c8-a443-3c1c3d73c458 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 748.541096] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.541290] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.541690] env[61991]: DEBUG nova.network.neutron [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.566030] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129567, 'name': CreateVM_Task, 'duration_secs': 0.387521} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.566236] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 748.566979] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.567189] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.567516] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.567807] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c2cd458-19b0-435d-a14b-8b26a48ab7d4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.573421] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 748.573421] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b5c392-3295-2485-ad2f-2c1d108def55" [ 748.573421] env[61991]: _type = "Task" [ 748.573421] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.584746] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b5c392-3295-2485-ad2f-2c1d108def55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.592098] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.689868] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129568, 'name': Rename_Task, 'duration_secs': 0.260796} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.693067] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 748.693918] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a2b057b-0f0d-49a9-b5e1-4dd5d021e467 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.701400] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 748.701400] env[61991]: value = "task-1129569" [ 748.701400] env[61991]: _type = "Task" [ 748.701400] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.716255] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.797087] env[61991]: DEBUG oslo_concurrency.lockutils [req-524b98d7-cc8a-4cce-8e12-0b3a743cbbec req-5c25ad1f-8a9f-4902-88d1-ca9c46de137a service nova] Releasing lock "refresh_cache-700ce4a7-d597-449a-9379-0cfb0c8f82fc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.935667] env[61991]: DEBUG nova.network.neutron [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Updated VIF entry in instance network info cache for port b1cd4427-17b1-40c2-8937-03d1f29e3d21. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 748.936161] env[61991]: DEBUG nova.network.neutron [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Updating instance_info_cache with network_info: [{"id": "b1cd4427-17b1-40c2-8937-03d1f29e3d21", "address": "fa:16:3e:17:5b:5b", "network": {"id": "334b8767-1629-42bc-94cc-70a52b88b1dc", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1266141268-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "53dfdbbaa1094205b865d3fcf591136e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1cd4427-17", "ovs_interfaceid": "b1cd4427-17b1-40c2-8937-03d1f29e3d21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.067482] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b940d5b-ef3a-45de-a2da-e10591a4da38 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.079351] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ef4be1-7f50-4bc3-96bf-0446f834e9a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.089091] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b5c392-3295-2485-ad2f-2c1d108def55, 'name': SearchDatastore_Task, 'duration_secs': 0.012707} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.119696] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.119812] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 749.119976] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.120148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.120398] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 749.124724] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ab5b3c5-52bf-40f2-ab7f-0eabe12934a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.130321] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e972bf17-9455-4e8f-bcbc-17860781897a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.140098] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b123e3e-86de-4da3-b46e-a95c0df25032 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.144421] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.144882] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 749.145654] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9903a8ee-b959-4c1c-804f-578c8b37eab1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.157685] env[61991]: DEBUG nova.compute.provider_tree [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.163444] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 749.163444] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b9f2d9-3afe-cdc9-fdcb-0a5958e0231b" [ 749.163444] env[61991]: _type = "Task" [ 749.163444] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.173408] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b9f2d9-3afe-cdc9-fdcb-0a5958e0231b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.214084] env[61991]: DEBUG oslo_vmware.api [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129569, 'name': PowerOnVM_Task, 'duration_secs': 0.507027} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.214718] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.214860] env[61991]: INFO nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Took 6.73 seconds to spawn the instance on the hypervisor. [ 749.214975] env[61991]: DEBUG nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.215804] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369f5f4f-b6e7-4bce-ba62-1ec7c4d48a1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.402705] env[61991]: DEBUG nova.network.neutron [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.442009] env[61991]: DEBUG oslo_concurrency.lockutils [req-8f002431-b58c-4e28-ab13-0e6e63beeec8 req-ca1e9f74-516a-4295-951a-b01174967c86 service nova] Releasing lock "refresh_cache-4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.635432] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.661178] env[61991]: DEBUG nova.scheduler.client.report [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.669656] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.670446] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.670741] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.671316] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.671569] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.671858] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.672193] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.672436] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.672725] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.673031] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.673240] env[61991]: DEBUG nova.virt.hardware [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.675754] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e05bcf0-07af-4984-bc9b-4f5e933bfd14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.693620] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6259333e-06de-40e1-8e36-a2218431c13e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.698165] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b9f2d9-3afe-cdc9-fdcb-0a5958e0231b, 'name': SearchDatastore_Task, 'duration_secs': 0.009907} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.699525] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8acc91b6-015e-4aac-896a-a3215c1adce0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.714220] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 749.714220] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c54647-d129-a865-f19e-e5ec012dc6ed" [ 749.714220] env[61991]: _type = "Task" [ 749.714220] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.721749] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c54647-d129-a865-f19e-e5ec012dc6ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.733069] env[61991]: INFO nova.compute.manager [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Took 37.30 seconds to build instance. [ 749.905639] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.180871] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.601s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.181448] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.184166] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.356s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.185787] env[61991]: INFO nova.compute.claims [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.225646] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c54647-d129-a865-f19e-e5ec012dc6ed, 'name': SearchDatastore_Task, 'duration_secs': 0.020406} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.229305] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.229305] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349/4b3c05a3-c40a-40c8-8501-5c1b8e4ea349.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 750.229305] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9db17bb6-9f11-4e4a-bbfb-cf4fe7ce795a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.236030] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 750.236030] env[61991]: value = "task-1129570" [ 750.236030] env[61991]: _type = "Task" [ 750.236030] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.242458] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c9681748-63f1-439e-a8ca-696d9ff3418c tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.526s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.249142] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129570, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.409736] env[61991]: DEBUG nova.compute.manager [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61991) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 750.410169] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.691042] env[61991]: DEBUG nova.compute.utils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.694965] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 750.695319] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 750.735396] env[61991]: DEBUG nova.compute.manager [req-3a594ae8-01ba-4ad3-8988-0d50b33be4d6 req-63899a59-eec1-413e-b18b-5c5889b97e7d service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Received event network-vif-plugged-248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.735396] env[61991]: DEBUG oslo_concurrency.lockutils [req-3a594ae8-01ba-4ad3-8988-0d50b33be4d6 req-63899a59-eec1-413e-b18b-5c5889b97e7d service nova] Acquiring lock "7df78da9-bf65-4621-b50d-43f1d721c2f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.736930] env[61991]: DEBUG oslo_concurrency.lockutils [req-3a594ae8-01ba-4ad3-8988-0d50b33be4d6 req-63899a59-eec1-413e-b18b-5c5889b97e7d service nova] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.736930] env[61991]: DEBUG oslo_concurrency.lockutils [req-3a594ae8-01ba-4ad3-8988-0d50b33be4d6 req-63899a59-eec1-413e-b18b-5c5889b97e7d service nova] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.737714] env[61991]: DEBUG nova.compute.manager [req-3a594ae8-01ba-4ad3-8988-0d50b33be4d6 req-63899a59-eec1-413e-b18b-5c5889b97e7d service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] No waiting events found dispatching network-vif-plugged-248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 750.738259] env[61991]: WARNING nova.compute.manager [req-3a594ae8-01ba-4ad3-8988-0d50b33be4d6 req-63899a59-eec1-413e-b18b-5c5889b97e7d service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Received unexpected event network-vif-plugged-248dc7ce-d07a-4030-810d-e49b969cb8ab for instance with vm_state building and task_state spawning. [ 750.746575] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.757734] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129570, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503062} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.757734] env[61991]: DEBUG nova.network.neutron [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Successfully updated port: 248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 750.761624] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349/4b3c05a3-c40a-40c8-8501-5c1b8e4ea349.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 750.761624] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.761952] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a5c0f9a-d9e3-4b28-a0a0-853091586763 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.773026] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 750.773026] env[61991]: value = "task-1129571" [ 750.773026] env[61991]: _type = "Task" [ 750.773026] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.783800] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.817318] env[61991]: DEBUG nova.compute.manager [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Received event network-changed-29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.817538] env[61991]: DEBUG nova.compute.manager [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Refreshing instance network info cache due to event network-changed-29738db6-7735-48b9-920a-e9d98d76aa24. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.817782] env[61991]: DEBUG oslo_concurrency.lockutils [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] Acquiring lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.818506] env[61991]: DEBUG oslo_concurrency.lockutils [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] Acquired lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.818506] env[61991]: DEBUG nova.network.neutron [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Refreshing network info cache for port 29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 750.823201] env[61991]: DEBUG nova.policy [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '081aa99dd10c4255ac2a663ed424440e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a44a2e26d9f45bcba0cdeecf09a04c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.203944] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.260823] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "refresh_cache-7df78da9-bf65-4621-b50d-43f1d721c2f1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.260981] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquired lock "refresh_cache-7df78da9-bf65-4621-b50d-43f1d721c2f1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.262043] env[61991]: DEBUG nova.network.neutron [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.279059] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.285377] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069636} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.288427] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.289778] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db94556-a74c-4bab-99a9-ed2b18d8e8c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.318014] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349/4b3c05a3-c40a-40c8-8501-5c1b8e4ea349.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.319390] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f230ccb-0076-4704-97a4-36b41778ed63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.350220] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 751.350220] env[61991]: value = "task-1129572" [ 751.350220] env[61991]: _type = "Task" [ 751.350220] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.361738] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129572, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.382503] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Successfully created port: 46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.483109] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "3e06e424-b64e-4fd2-8013-27760200c41d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.483376] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.483608] env[61991]: INFO nova.compute.manager [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Rebooting instance [ 751.627932] env[61991]: DEBUG nova.network.neutron [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updated VIF entry in instance network info cache for port 29738db6-7735-48b9-920a-e9d98d76aa24. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 751.627932] env[61991]: DEBUG nova.network.neutron [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updating instance_info_cache with network_info: [{"id": "29738db6-7735-48b9-920a-e9d98d76aa24", "address": "fa:16:3e:67:8e:c4", "network": {"id": "7bdca730-2d82-430c-9297-63bf61dec70e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1406357972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ddab761ff5f49aeb8a4b611cd9d7603", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29738db6-77", "ovs_interfaceid": "29738db6-7735-48b9-920a-e9d98d76aa24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.654797] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57162f6-49db-4091-ad4b-a2efd4409ba0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.663871] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a1294b-d2d0-424a-b5d8-a528e759d2c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.698366] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a225aa63-5c0f-4f46-943b-eff1a3fdafc1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.706727] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdc4949-68a2-4f27-b27a-c8b92ae9339d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.726724] env[61991]: DEBUG nova.compute.provider_tree [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.813814] env[61991]: DEBUG nova.network.neutron [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.861306] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129572, 'name': ReconfigVM_Task, 'duration_secs': 0.27162} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.861306] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349/4b3c05a3-c40a-40c8-8501-5c1b8e4ea349.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 751.861306] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0251ff4a-8da1-425c-9c6c-e390dc02666f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.870773] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 751.870773] env[61991]: value = "task-1129573" [ 751.870773] env[61991]: _type = "Task" [ 751.870773] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.884647] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129573, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.029342] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.041365] env[61991]: DEBUG nova.network.neutron [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Updating instance_info_cache with network_info: [{"id": "248dc7ce-d07a-4030-810d-e49b969cb8ab", "address": "fa:16:3e:7b:66:22", "network": {"id": "5057485d-9068-4aae-a3d8-a8a2ee4e9c28", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-141779178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a0346ea1b8f4cf9a9322dea401e7b43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap248dc7ce-d0", "ovs_interfaceid": "248dc7ce-d07a-4030-810d-e49b969cb8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.133732] env[61991]: DEBUG oslo_concurrency.lockutils [req-41f02243-dd85-459e-98f6-0d988fdf53df req-6abd0a01-6b60-49b4-abf4-aa8015886cdc service nova] Releasing lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.134265] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquired lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.134473] env[61991]: DEBUG nova.network.neutron [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.217292] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.230649] env[61991]: DEBUG nova.scheduler.client.report [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.250733] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.251055] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.251243] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.251451] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.251621] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.251780] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.251994] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.252652] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.252854] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.253062] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.253270] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.254192] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de159d12-0827-4c3c-867d-36afe8945ba6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.266663] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b20e680-2638-4c3c-84c0-82d9da311d81 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.385413] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129573, 'name': Rename_Task, 'duration_secs': 0.156678} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.385413] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.385413] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32b86944-ae88-47ae-97f9-7ffa16a945a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.391164] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 752.391164] env[61991]: value = "task-1129574" [ 752.391164] env[61991]: _type = "Task" [ 752.391164] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.402660] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129574, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.544077] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Releasing lock "refresh_cache-7df78da9-bf65-4621-b50d-43f1d721c2f1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.544481] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Instance network_info: |[{"id": "248dc7ce-d07a-4030-810d-e49b969cb8ab", "address": "fa:16:3e:7b:66:22", "network": {"id": "5057485d-9068-4aae-a3d8-a8a2ee4e9c28", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-141779178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a0346ea1b8f4cf9a9322dea401e7b43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap248dc7ce-d0", "ovs_interfaceid": "248dc7ce-d07a-4030-810d-e49b969cb8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 752.544983] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:66:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ded18042-834c-4792-b3e8-b1c377446432', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '248dc7ce-d07a-4030-810d-e49b969cb8ab', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.555130] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Creating folder: Project (0a0346ea1b8f4cf9a9322dea401e7b43). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.555410] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a25d020-1c15-4b83-af67-dad651b4c963 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.565647] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Created folder: Project (0a0346ea1b8f4cf9a9322dea401e7b43) in parent group-v246753. [ 752.565849] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Creating folder: Instances. Parent ref: group-v246842. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.566107] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f040f290-7217-493e-afc0-c75169baab30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.576509] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Created folder: Instances in parent group-v246842. [ 752.576509] env[61991]: DEBUG oslo.service.loopingcall [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.576509] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.576509] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6020f726-854a-4672-9aac-fe540b2b3010 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.597092] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.597092] env[61991]: value = "task-1129577" [ 752.597092] env[61991]: _type = "Task" [ 752.597092] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.605846] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129577, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.734932] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.735510] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 752.738283] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.322s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.738367] env[61991]: DEBUG nova.objects.instance [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lazy-loading 'resources' on Instance uuid 0631634f-5c58-4ab0-b0dd-71821caf2fc6 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 752.772063] env[61991]: DEBUG nova.compute.manager [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Received event network-changed-248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.772063] env[61991]: DEBUG nova.compute.manager [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Refreshing instance network info cache due to event network-changed-248dc7ce-d07a-4030-810d-e49b969cb8ab. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 752.772374] env[61991]: DEBUG oslo_concurrency.lockutils [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] Acquiring lock "refresh_cache-7df78da9-bf65-4621-b50d-43f1d721c2f1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.772374] env[61991]: DEBUG oslo_concurrency.lockutils [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] Acquired lock "refresh_cache-7df78da9-bf65-4621-b50d-43f1d721c2f1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.772567] env[61991]: DEBUG nova.network.neutron [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Refreshing network info cache for port 248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.907417] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129574, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.941890] env[61991]: DEBUG nova.network.neutron [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updating instance_info_cache with network_info: [{"id": "29738db6-7735-48b9-920a-e9d98d76aa24", "address": "fa:16:3e:67:8e:c4", "network": {"id": "7bdca730-2d82-430c-9297-63bf61dec70e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1406357972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ddab761ff5f49aeb8a4b611cd9d7603", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29738db6-77", "ovs_interfaceid": "29738db6-7735-48b9-920a-e9d98d76aa24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.109746] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129577, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.241161] env[61991]: DEBUG nova.compute.utils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.250453] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 753.251145] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 753.273788] env[61991]: DEBUG nova.compute.manager [req-7178a179-85eb-43a6-9e7b-f2280076cdf5 req-3956d092-cc37-4bad-bffc-b9da07c28a72 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Received event network-vif-plugged-46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.273788] env[61991]: DEBUG oslo_concurrency.lockutils [req-7178a179-85eb-43a6-9e7b-f2280076cdf5 req-3956d092-cc37-4bad-bffc-b9da07c28a72 service nova] Acquiring lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.273788] env[61991]: DEBUG oslo_concurrency.lockutils [req-7178a179-85eb-43a6-9e7b-f2280076cdf5 req-3956d092-cc37-4bad-bffc-b9da07c28a72 service nova] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.273788] env[61991]: DEBUG oslo_concurrency.lockutils [req-7178a179-85eb-43a6-9e7b-f2280076cdf5 req-3956d092-cc37-4bad-bffc-b9da07c28a72 service nova] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.273788] env[61991]: DEBUG nova.compute.manager [req-7178a179-85eb-43a6-9e7b-f2280076cdf5 req-3956d092-cc37-4bad-bffc-b9da07c28a72 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] No waiting events found dispatching network-vif-plugged-46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 753.276875] env[61991]: WARNING nova.compute.manager [req-7178a179-85eb-43a6-9e7b-f2280076cdf5 req-3956d092-cc37-4bad-bffc-b9da07c28a72 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Received unexpected event network-vif-plugged-46746327-a6c2-4236-b0eb-823d21d735a4 for instance with vm_state building and task_state spawning. [ 753.379556] env[61991]: DEBUG nova.policy [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '081aa99dd10c4255ac2a663ed424440e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a44a2e26d9f45bcba0cdeecf09a04c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.384137] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Successfully updated port: 46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 753.406642] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129574, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.445851] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Releasing lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.448781] env[61991]: DEBUG nova.compute.manager [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.449752] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f983abf5-b337-49d7-be8e-4b2b4742e4b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.500565] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.500807] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.614744] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129577, 'name': CreateVM_Task, 'duration_secs': 0.711577} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.619082] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 753.622073] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.622687] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.622687] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 753.622938] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aea3839-6af0-47e0-8683-804ff6abc854 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.628311] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 753.628311] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286b657-9443-a8cd-5d29-47f8fbdf6478" [ 753.628311] env[61991]: _type = "Task" [ 753.628311] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.638797] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286b657-9443-a8cd-5d29-47f8fbdf6478, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.643588] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.645022] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.738750] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0e69a6-107d-4887-88a9-3078d95e4b2c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.747160] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9641eb2f-3596-4f4d-94e8-3f93f9988a21 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.750959] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 753.781352] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ec981c-6b46-4a54-9d84-1505df0568d3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.790023] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dded340-fd40-4b40-b1e6-7af2d7a3414e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.803610] env[61991]: DEBUG nova.compute.provider_tree [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.809895] env[61991]: DEBUG nova.network.neutron [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Updated VIF entry in instance network info cache for port 248dc7ce-d07a-4030-810d-e49b969cb8ab. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 753.810398] env[61991]: DEBUG nova.network.neutron [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Updating instance_info_cache with network_info: [{"id": "248dc7ce-d07a-4030-810d-e49b969cb8ab", "address": "fa:16:3e:7b:66:22", "network": {"id": "5057485d-9068-4aae-a3d8-a8a2ee4e9c28", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-141779178-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a0346ea1b8f4cf9a9322dea401e7b43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap248dc7ce-d0", "ovs_interfaceid": "248dc7ce-d07a-4030-810d-e49b969cb8ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.886200] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "refresh_cache-d2a0e5d1-5557-41b6-936d-dc86d1346c61" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.886511] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "refresh_cache-d2a0e5d1-5557-41b6-936d-dc86d1346c61" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.886684] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.909585] env[61991]: DEBUG oslo_vmware.api [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129574, 'name': PowerOnVM_Task, 'duration_secs': 1.339876} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.910827] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.911203] env[61991]: INFO nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Took 8.90 seconds to spawn the instance on the hypervisor. [ 753.911530] env[61991]: DEBUG nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.912897] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c024050c-6813-4fda-852a-ceeaeb34d1a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.947930] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Successfully created port: 790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.140973] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5286b657-9443-a8cd-5d29-47f8fbdf6478, 'name': SearchDatastore_Task, 'duration_secs': 0.01238} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.141608] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.142023] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.142427] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.142731] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.143093] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.143468] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bed811b1-164d-437f-99f1-5404f59e0082 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.146973] env[61991]: DEBUG nova.compute.utils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.154205] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.154655] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 754.155597] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5576827-99d0-469f-9ac5-7c63913e7bc9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.161896] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 754.161896] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5248ed63-8410-42ea-c24a-e0ba1561912e" [ 754.161896] env[61991]: _type = "Task" [ 754.161896] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.172611] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5248ed63-8410-42ea-c24a-e0ba1561912e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.308263] env[61991]: DEBUG nova.scheduler.client.report [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.314444] env[61991]: DEBUG oslo_concurrency.lockutils [req-5a7a680e-aa13-485c-b7af-abd6b4eae653 req-06b24876-fa7b-404c-b086-88c6988f979c service nova] Releasing lock "refresh_cache-7df78da9-bf65-4621-b50d-43f1d721c2f1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.435663] env[61991]: INFO nova.compute.manager [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Took 39.52 seconds to build instance. [ 754.445158] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.471451] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64104a91-ef50-42fd-9647-df4eaba7743f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.482513] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Doing hard reboot of VM {{(pid=61991) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 754.482764] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-90e5ed21-1a35-444a-a4cb-69071434005c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.489464] env[61991]: DEBUG oslo_vmware.api [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 754.489464] env[61991]: value = "task-1129578" [ 754.489464] env[61991]: _type = "Task" [ 754.489464] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.497971] env[61991]: DEBUG oslo_vmware.api [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129578, 'name': ResetVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.649746] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.672567] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5248ed63-8410-42ea-c24a-e0ba1561912e, 'name': SearchDatastore_Task, 'duration_secs': 0.015664} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.673438] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9a9c8d5-9dc0-464b-9812-175fd7a90b54 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.679845] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 754.679845] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fa9416-6d35-af1e-fceb-2a4b182833c3" [ 754.679845] env[61991]: _type = "Task" [ 754.679845] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.683676] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Updating instance_info_cache with network_info: [{"id": "46746327-a6c2-4236-b0eb-823d21d735a4", "address": "fa:16:3e:36:0d:27", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46746327-a6", "ovs_interfaceid": "46746327-a6c2-4236-b0eb-823d21d735a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.690640] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fa9416-6d35-af1e-fceb-2a4b182833c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.762974] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.789632] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.789911] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.790275] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.791029] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.791029] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.791029] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.791232] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.791658] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.791658] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.791758] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.791868] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.792871] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbda861-bec2-43d2-bffb-a7b7c63c4e4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.801083] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a38385b-5b9f-4d0b-a344-674c30eaa0c9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.817080] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.079s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.820261] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.921s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.821981] env[61991]: INFO nova.compute.claims [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.840394] env[61991]: INFO nova.scheduler.client.report [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Deleted allocations for instance 0631634f-5c58-4ab0-b0dd-71821caf2fc6 [ 754.938597] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9fe4e66c-0f33-430d-ba04-756963fb9588 tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.707s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.999557] env[61991]: DEBUG oslo_vmware.api [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129578, 'name': ResetVM_Task, 'duration_secs': 0.119079} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.999834] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Did hard reboot of VM {{(pid=61991) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 755.000027] env[61991]: DEBUG nova.compute.manager [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.001022] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94972c94-65f6-4d96-bf8d-ce2e43c80afc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.191414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "refresh_cache-d2a0e5d1-5557-41b6-936d-dc86d1346c61" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.191414] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Instance network_info: |[{"id": "46746327-a6c2-4236-b0eb-823d21d735a4", "address": "fa:16:3e:36:0d:27", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46746327-a6", "ovs_interfaceid": "46746327-a6c2-4236-b0eb-823d21d735a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 755.191414] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fa9416-6d35-af1e-fceb-2a4b182833c3, 'name': SearchDatastore_Task, 'duration_secs': 0.016778} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.191828] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:0d:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46746327-a6c2-4236-b0eb-823d21d735a4', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.200011] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Creating folder: Project (5a44a2e26d9f45bcba0cdeecf09a04c8). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.200448] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.200831] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7df78da9-bf65-4621-b50d-43f1d721c2f1/7df78da9-bf65-4621-b50d-43f1d721c2f1.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 755.201123] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-742a4af4-84e2-4a71-a56f-bc02fac10617 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.203034] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-757924e1-f1a7-4f91-bf66-8224940e27cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.210667] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 755.210667] env[61991]: value = "task-1129580" [ 755.210667] env[61991]: _type = "Task" [ 755.210667] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.223593] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Created folder: Project (5a44a2e26d9f45bcba0cdeecf09a04c8) in parent group-v246753. [ 755.223941] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Creating folder: Instances. Parent ref: group-v246845. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.224991] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9e1565a-5b7f-4bf8-b800-24bacb418d95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.233582] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129580, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.235567] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Created folder: Instances in parent group-v246845. [ 755.235813] env[61991]: DEBUG oslo.service.loopingcall [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.236017] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.236237] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72aae422-3e00-4791-8c03-3de0798b0f28 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.255660] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.255660] env[61991]: value = "task-1129582" [ 755.255660] env[61991]: _type = "Task" [ 755.255660] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.267707] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129582, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.348386] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fa6496b9-7cfc-433c-8ac1-be03426049a7 tempest-ServersTestMultiNic-1612236620 tempest-ServersTestMultiNic-1612236620-project-member] Lock "0631634f-5c58-4ab0-b0dd-71821caf2fc6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.330s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.434539] env[61991]: DEBUG nova.compute.manager [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Received event network-changed-46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.434539] env[61991]: DEBUG nova.compute.manager [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Refreshing instance network info cache due to event network-changed-46746327-a6c2-4236-b0eb-823d21d735a4. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 755.434539] env[61991]: DEBUG oslo_concurrency.lockutils [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] Acquiring lock "refresh_cache-d2a0e5d1-5557-41b6-936d-dc86d1346c61" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.434539] env[61991]: DEBUG oslo_concurrency.lockutils [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] Acquired lock "refresh_cache-d2a0e5d1-5557-41b6-936d-dc86d1346c61" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.434689] env[61991]: DEBUG nova.network.neutron [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Refreshing network info cache for port 46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 755.442028] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.514417] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e4998cde-45b6-4e64-99f5-726d1b6c44b3 tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.030s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.668189] env[61991]: DEBUG nova.compute.manager [req-2666cd8e-29d5-41f3-868c-ef893db05404 req-05fd6d48-5592-45e9-9848-88747138eb63 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Received event network-vif-plugged-790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.668863] env[61991]: DEBUG oslo_concurrency.lockutils [req-2666cd8e-29d5-41f3-868c-ef893db05404 req-05fd6d48-5592-45e9-9848-88747138eb63 service nova] Acquiring lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.668863] env[61991]: DEBUG oslo_concurrency.lockutils [req-2666cd8e-29d5-41f3-868c-ef893db05404 req-05fd6d48-5592-45e9-9848-88747138eb63 service nova] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.668863] env[61991]: DEBUG oslo_concurrency.lockutils [req-2666cd8e-29d5-41f3-868c-ef893db05404 req-05fd6d48-5592-45e9-9848-88747138eb63 service nova] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.669136] env[61991]: DEBUG nova.compute.manager [req-2666cd8e-29d5-41f3-868c-ef893db05404 req-05fd6d48-5592-45e9-9848-88747138eb63 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] No waiting events found dispatching network-vif-plugged-790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 755.669249] env[61991]: WARNING nova.compute.manager [req-2666cd8e-29d5-41f3-868c-ef893db05404 req-05fd6d48-5592-45e9-9848-88747138eb63 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Received unexpected event network-vif-plugged-790d5666-4e4c-4dcf-8f02-6a40df7a7666 for instance with vm_state building and task_state spawning. [ 755.727295] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129580, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.729288] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Successfully updated port: 790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 755.730907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.730907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.730907] env[61991]: INFO nova.compute.manager [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Attaching volume 48ca7de5-1564-44be-83d4-3298210a3c5d to /dev/sdb [ 755.776070] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129582, 'name': CreateVM_Task, 'duration_secs': 0.418595} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.776357] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.778306] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.778651] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.779490] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.781044] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2624ada-fc6a-49f9-b795-cc25c4afcce8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.784530] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efaaeb4b-5d38-4fca-a628-9493094ce37e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.793407] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 755.793407] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52365f1e-7594-bbca-8a35-7beb2ce3ac4e" [ 755.793407] env[61991]: _type = "Task" [ 755.793407] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.801683] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589c4614-f52f-4cf0-b28a-2eac2a8be75f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.808668] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52365f1e-7594-bbca-8a35-7beb2ce3ac4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.824384] env[61991]: DEBUG nova.virt.block_device [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating existing volume attachment record: 9aa147f2-c501-41f0-a249-c496e0f8f844 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 755.972232] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.125385] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286f284-1aa3-ed90-8861-476295d8a854/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 756.126694] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c85cec-deb7-4af9-a64b-eddfbdd82d29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.136870] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286f284-1aa3-ed90-8861-476295d8a854/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 756.137091] env[61991]: ERROR oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286f284-1aa3-ed90-8861-476295d8a854/disk-0.vmdk due to incomplete transfer. [ 756.137338] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e6de5f67-b2dc-4286-bdc6-0bb42ebbaab4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.149547] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286f284-1aa3-ed90-8861-476295d8a854/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 756.149767] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Uploaded image e8cac67d-5662-46ad-be70-bc504bd13ab9 to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 756.152430] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 756.152430] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7c274c93-fa53-4874-a9fb-f1ed5bfac067 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.158714] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 756.158714] env[61991]: value = "task-1129585" [ 756.158714] env[61991]: _type = "Task" [ 756.158714] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.169384] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129585, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.222811] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129580, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691025} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.223197] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7df78da9-bf65-4621-b50d-43f1d721c2f1/7df78da9-bf65-4621-b50d-43f1d721c2f1.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 756.223454] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.223699] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62a12446-2b2d-4b77-8824-33a59b25f330 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.232035] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 756.232035] env[61991]: value = "task-1129587" [ 756.232035] env[61991]: _type = "Task" [ 756.232035] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.237090] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "refresh_cache-7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.237090] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "refresh_cache-7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.237090] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 756.242460] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.273324] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a403818b-13c4-4dfd-94f0-251636781378 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.287806] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2077e1-d325-43c4-97bc-3f9af29582e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.330966] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d230b6-5dfe-4fbf-b4d1-c2edb53de371 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.343054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee04aa8-2bb5-4574-afc3-92a00bc37643 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.347117] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52365f1e-7594-bbca-8a35-7beb2ce3ac4e, 'name': SearchDatastore_Task, 'duration_secs': 0.070869} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.347458] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.347833] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.347977] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.348282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.348366] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.349115] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13625957-a21a-4fc6-bb7f-0f91a69e9938 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.363156] env[61991]: DEBUG nova.compute.provider_tree [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.370456] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.370858] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 756.372405] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ee73cd2-bb3f-434b-8330-8948120c3741 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.378857] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 756.378857] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c1810e-ae45-cd7d-9d9e-5e1ec0da4689" [ 756.378857] env[61991]: _type = "Task" [ 756.378857] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.388223] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c1810e-ae45-cd7d-9d9e-5e1ec0da4689, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.389353] env[61991]: DEBUG nova.network.neutron [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Updated VIF entry in instance network info cache for port 46746327-a6c2-4236-b0eb-823d21d735a4. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 756.389708] env[61991]: DEBUG nova.network.neutron [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Updating instance_info_cache with network_info: [{"id": "46746327-a6c2-4236-b0eb-823d21d735a4", "address": "fa:16:3e:36:0d:27", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46746327-a6", "ovs_interfaceid": "46746327-a6c2-4236-b0eb-823d21d735a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.673913] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129585, 'name': Destroy_Task, 'duration_secs': 0.371433} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.674507] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Destroyed the VM [ 756.674917] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 756.675257] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5eacb2b6-d6f7-4dae-9626-7654b68c80a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.682916] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 756.682916] env[61991]: value = "task-1129588" [ 756.682916] env[61991]: _type = "Task" [ 756.682916] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.695920] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129588, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.743094] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087612} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.745381] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.746680] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56e29da-9270-471c-b8d3-a7218faacec3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.772805] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 7df78da9-bf65-4621-b50d-43f1d721c2f1/7df78da9-bf65-4621-b50d-43f1d721c2f1.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.773360] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2237195f-f8d0-462d-8328-a3d03dbafae9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.791806] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.798882] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 756.798882] env[61991]: value = "task-1129589" [ 756.798882] env[61991]: _type = "Task" [ 756.798882] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.811080] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129589, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.865536] env[61991]: DEBUG nova.scheduler.client.report [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.871224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "49324fef-ad48-451b-a5ce-d9a1231137db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.871595] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.889734] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c1810e-ae45-cd7d-9d9e-5e1ec0da4689, 'name': SearchDatastore_Task, 'duration_secs': 0.022812} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.891089] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10998558-4a56-415a-9aa4-e136a6f337f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.896263] env[61991]: DEBUG oslo_concurrency.lockutils [req-3e66c501-c7c0-4488-a2d7-73811216df70 req-daab680d-7a1b-4610-9f23-af1f7a2760f0 service nova] Releasing lock "refresh_cache-d2a0e5d1-5557-41b6-936d-dc86d1346c61" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.900062] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 756.900062] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5242a954-a37a-6eb0-2a0c-dec04949800f" [ 756.900062] env[61991]: _type = "Task" [ 756.900062] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.908379] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5242a954-a37a-6eb0-2a0c-dec04949800f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.986507] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Updating instance_info_cache with network_info: [{"id": "790d5666-4e4c-4dcf-8f02-6a40df7a7666", "address": "fa:16:3e:ad:3a:0a", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap790d5666-4e", "ovs_interfaceid": "790d5666-4e4c-4dcf-8f02-6a40df7a7666", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.193346] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129588, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.309589] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.372955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.373530] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.376820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.866s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.378167] env[61991]: INFO nova.compute.claims [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.416970] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5242a954-a37a-6eb0-2a0c-dec04949800f, 'name': SearchDatastore_Task, 'duration_secs': 0.034372} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.416970] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.416970] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] d2a0e5d1-5557-41b6-936d-dc86d1346c61/d2a0e5d1-5557-41b6-936d-dc86d1346c61.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 757.416970] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fbc27c5-dc34-4326-8cd3-7dfdb288ad35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.422979] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 757.422979] env[61991]: value = "task-1129590" [ 757.422979] env[61991]: _type = "Task" [ 757.422979] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.434623] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129590, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.447122] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.447371] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.447582] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.447765] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.447939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.450039] env[61991]: INFO nova.compute.manager [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Terminating instance [ 757.452185] env[61991]: DEBUG nova.compute.manager [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.452414] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.453304] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e37573-985e-4690-a7a4-c95fa150db66 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.461324] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 757.461324] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-778df31c-0fab-472a-b611-2de717ecedab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.469827] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 757.469827] env[61991]: value = "task-1129591" [ 757.469827] env[61991]: _type = "Task" [ 757.469827] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.480320] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.488110] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "refresh_cache-7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.489131] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Instance network_info: |[{"id": "790d5666-4e4c-4dcf-8f02-6a40df7a7666", "address": "fa:16:3e:ad:3a:0a", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap790d5666-4e", "ovs_interfaceid": "790d5666-4e4c-4dcf-8f02-6a40df7a7666", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 757.489131] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:3a:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '790d5666-4e4c-4dcf-8f02-6a40df7a7666', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.497581] env[61991]: DEBUG oslo.service.loopingcall [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.497904] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.498222] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08df6c25-a97d-4323-83dd-69cbaf6974e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.522610] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.522610] env[61991]: value = "task-1129592" [ 757.522610] env[61991]: _type = "Task" [ 757.522610] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.531447] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129592, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.684291] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "3e06e424-b64e-4fd2-8013-27760200c41d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.684537] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.684764] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "3e06e424-b64e-4fd2-8013-27760200c41d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.684954] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.685168] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.690783] env[61991]: INFO nova.compute.manager [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Terminating instance [ 757.693397] env[61991]: DEBUG nova.compute.manager [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.693614] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.694508] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb5b183-4010-4d22-b881-89077e312cea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.700830] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129588, 'name': RemoveSnapshot_Task, 'duration_secs': 0.863088} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.703306] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 757.703306] env[61991]: DEBUG nova.compute.manager [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.704893] env[61991]: DEBUG nova.compute.manager [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Received event network-changed-790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.705170] env[61991]: DEBUG nova.compute.manager [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Refreshing instance network info cache due to event network-changed-790d5666-4e4c-4dcf-8f02-6a40df7a7666. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.705340] env[61991]: DEBUG oslo_concurrency.lockutils [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] Acquiring lock "refresh_cache-7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.705468] env[61991]: DEBUG oslo_concurrency.lockutils [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] Acquired lock "refresh_cache-7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.705642] env[61991]: DEBUG nova.network.neutron [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Refreshing network info cache for port 790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.709884] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95dd65e1-5fc8-4b70-a045-3f6484ee722e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.713067] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 757.714218] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1191575-cfff-45cd-bf91-4c0b072deced {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.721249] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 757.721249] env[61991]: value = "task-1129593" [ 757.721249] env[61991]: _type = "Task" [ 757.721249] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.739095] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129593, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.813326] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129589, 'name': ReconfigVM_Task, 'duration_secs': 0.906659} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.813738] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 7df78da9-bf65-4621-b50d-43f1d721c2f1/7df78da9-bf65-4621-b50d-43f1d721c2f1.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.814467] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dfd36ff1-775c-499d-84df-f7b3ec595344 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.821591] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 757.821591] env[61991]: value = "task-1129594" [ 757.821591] env[61991]: _type = "Task" [ 757.821591] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.832214] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129594, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.890708] env[61991]: DEBUG nova.compute.utils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.892457] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.892645] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 757.936814] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129590, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.975614] env[61991]: DEBUG nova.policy [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '081aa99dd10c4255ac2a663ed424440e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a44a2e26d9f45bcba0cdeecf09a04c8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.983901] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129591, 'name': PowerOffVM_Task, 'duration_secs': 0.263586} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.984333] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 757.984536] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 757.984865] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18259904-6127-4172-90f3-e268abfcdf5f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.035176] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129592, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.050916] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.051199] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.051389] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Deleting the datastore file [datastore2] 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.051652] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ae5d9ce-db2d-449d-b7ee-e6931abce167 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.058487] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for the task: (returnval){ [ 758.058487] env[61991]: value = "task-1129596" [ 758.058487] env[61991]: _type = "Task" [ 758.058487] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.070797] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.233208] env[61991]: INFO nova.compute.manager [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Shelve offloading [ 758.235826] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.241399] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c282f8c-85f6-4b5c-b8d1-188b4c697649 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.247024] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129593, 'name': PowerOffVM_Task, 'duration_secs': 0.273706} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.247024] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 758.247024] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 758.247024] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbdaa5df-39e9-4125-8295-3a2269bb07db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.250189] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 758.250189] env[61991]: value = "task-1129598" [ 758.250189] env[61991]: _type = "Task" [ 758.250189] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.260974] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.301900] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 758.302174] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 758.302373] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Deleting the datastore file [datastore2] 3e06e424-b64e-4fd2-8013-27760200c41d {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 758.302680] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc882f0b-b923-442e-b4e5-f4473c2b41b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.310221] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for the task: (returnval){ [ 758.310221] env[61991]: value = "task-1129600" [ 758.310221] env[61991]: _type = "Task" [ 758.310221] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.320918] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.330562] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129594, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.399044] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.417540] env[61991]: DEBUG nova.scheduler.client.report [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 758.433191] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129590, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.792676} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.433508] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] d2a0e5d1-5557-41b6-936d-dc86d1346c61/d2a0e5d1-5557-41b6-936d-dc86d1346c61.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 758.433782] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.434486] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7953b092-69a1-4ea3-8418-a0ecaca422c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.442047] env[61991]: DEBUG nova.scheduler.client.report [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 758.442047] env[61991]: DEBUG nova.compute.provider_tree [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 758.443088] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 758.443088] env[61991]: value = "task-1129601" [ 758.443088] env[61991]: _type = "Task" [ 758.443088] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.452429] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129601, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.453496] env[61991]: DEBUG nova.scheduler.client.report [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 758.473355] env[61991]: DEBUG nova.scheduler.client.report [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 758.532907] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129592, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.571445] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.736807] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Successfully created port: 35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.767352] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 758.767767] env[61991]: DEBUG nova.compute.manager [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 758.769141] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc1695d-1ae5-457d-845e-c7c159cdf491 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.780551] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.780551] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.780551] env[61991]: DEBUG nova.network.neutron [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.797304] env[61991]: DEBUG nova.network.neutron [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Updated VIF entry in instance network info cache for port 790d5666-4e4c-4dcf-8f02-6a40df7a7666. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 758.797304] env[61991]: DEBUG nova.network.neutron [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Updating instance_info_cache with network_info: [{"id": "790d5666-4e4c-4dcf-8f02-6a40df7a7666", "address": "fa:16:3e:ad:3a:0a", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap790d5666-4e", "ovs_interfaceid": "790d5666-4e4c-4dcf-8f02-6a40df7a7666", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.827868] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.842567] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129594, 'name': Rename_Task, 'duration_secs': 0.529374} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.842567] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.842567] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-437abd62-ee81-46d8-849f-1aeff4234be9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.850422] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 758.850422] env[61991]: value = "task-1129602" [ 758.850422] env[61991]: _type = "Task" [ 758.850422] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.863592] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.958134] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.471366} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.958310] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.959146] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5622655d-9705-4e36-989f-345d6bf99533 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.986846] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] d2a0e5d1-5557-41b6-936d-dc86d1346c61/d2a0e5d1-5557-41b6-936d-dc86d1346c61.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.990209] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95616231-9588-49c4-9031-59cc101c5563 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.012443] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 759.012443] env[61991]: value = "task-1129603" [ 759.012443] env[61991]: _type = "Task" [ 759.012443] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.023467] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129603, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.032624] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129592, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.071609] env[61991]: DEBUG oslo_vmware.api [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Task: {'id': task-1129596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.629072} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.075203] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.075838] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.075838] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.076019] env[61991]: INFO nova.compute.manager [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Took 1.62 seconds to destroy the instance on the hypervisor. [ 759.076271] env[61991]: DEBUG oslo.service.loopingcall [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.076935] env[61991]: DEBUG nova.compute.manager [-] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.077046] env[61991]: DEBUG nova.network.neutron [-] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.139734] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98c7701-b995-4716-9232-dc6bc2549cde {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.150459] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07b99cd-e7d9-4e3c-883e-c5bb8335936b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.186263] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a4050e-2ce0-4d70-9d98-cb52a1285857 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.195401] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7c0067-2840-43f3-8fd6-e3a59ec6354a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.214846] env[61991]: DEBUG nova.compute.provider_tree [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 759.299640] env[61991]: DEBUG oslo_concurrency.lockutils [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] Releasing lock "refresh_cache-7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.299906] env[61991]: DEBUG nova.compute.manager [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Received event network-changed-29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.300102] env[61991]: DEBUG nova.compute.manager [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Refreshing instance network info cache due to event network-changed-29738db6-7735-48b9-920a-e9d98d76aa24. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.300333] env[61991]: DEBUG oslo_concurrency.lockutils [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] Acquiring lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.300757] env[61991]: DEBUG oslo_concurrency.lockutils [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] Acquired lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.301035] env[61991]: DEBUG nova.network.neutron [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Refreshing network info cache for port 29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.323148] env[61991]: DEBUG oslo_vmware.api [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Task: {'id': task-1129600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.812235} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.323745] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.323903] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 759.324738] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 759.324738] env[61991]: INFO nova.compute.manager [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Took 1.63 seconds to destroy the instance on the hypervisor. [ 759.324738] env[61991]: DEBUG oslo.service.loopingcall [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.324862] env[61991]: DEBUG nova.compute.manager [-] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.324999] env[61991]: DEBUG nova.network.neutron [-] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.364913] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129602, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.423469] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.462072] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.462493] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.462769] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.462994] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.463179] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.463341] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.463562] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.464702] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.464702] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.464702] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.464702] env[61991]: DEBUG nova.virt.hardware [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.465791] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e960d4-aafc-494f-a59c-ea8bce0f0357 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.475614] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6f8b31-9a9a-4655-923c-c11897de0d92 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.525831] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129603, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.535363] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129592, 'name': CreateVM_Task, 'duration_secs': 1.822146} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.535363] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 759.536202] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.536202] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.536703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 759.536795] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd4d258e-effa-455c-996f-c931491a8d1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.541672] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 759.541672] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527965e6-6291-29cb-2351-cdfc85e49a08" [ 759.541672] env[61991]: _type = "Task" [ 759.541672] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.557223] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527965e6-6291-29cb-2351-cdfc85e49a08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.756191] env[61991]: DEBUG nova.compute.manager [req-9316df04-23db-4384-aed4-b9581a23b566 req-ac7741f4-97a2-4b95-90d5-4982cc9dcda1 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Received event network-vif-deleted-b1cd4427-17b1-40c2-8937-03d1f29e3d21 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.756356] env[61991]: INFO nova.compute.manager [req-9316df04-23db-4384-aed4-b9581a23b566 req-ac7741f4-97a2-4b95-90d5-4982cc9dcda1 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Neutron deleted interface b1cd4427-17b1-40c2-8937-03d1f29e3d21; detaching it from the instance and deleting it from the info cache [ 759.756594] env[61991]: DEBUG nova.network.neutron [req-9316df04-23db-4384-aed4-b9581a23b566 req-ac7741f4-97a2-4b95-90d5-4982cc9dcda1 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.764184] env[61991]: DEBUG nova.scheduler.client.report [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 759.764499] env[61991]: DEBUG nova.compute.provider_tree [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 77 to 78 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 759.764798] env[61991]: DEBUG nova.compute.provider_tree [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 759.827035] env[61991]: DEBUG nova.network.neutron [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Updating instance_info_cache with network_info: [{"id": "8aa50c96-062e-4d6c-9aa0-990a50838237", "address": "fa:16:3e:62:a0:f9", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aa50c96-06", "ovs_interfaceid": "8aa50c96-062e-4d6c-9aa0-990a50838237", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.836206] env[61991]: INFO nova.network.neutron [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Port 29738db6-7735-48b9-920a-e9d98d76aa24 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 759.836358] env[61991]: DEBUG nova.network.neutron [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.865959] env[61991]: DEBUG oslo_vmware.api [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129602, 'name': PowerOnVM_Task, 'duration_secs': 0.834823} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.866889] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.866889] env[61991]: INFO nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Took 10.23 seconds to spawn the instance on the hypervisor. [ 759.866889] env[61991]: DEBUG nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 759.867658] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d0dffb-ba48-4366-b0ce-75a2d75548b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.876525] env[61991]: DEBUG nova.compute.manager [req-3002b0a8-e1c8-4e13-82c2-4c3dad6d4a0f req-5468095a-b1ee-43bf-aa16-79cc4cae8bd3 service nova] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Received event network-vif-deleted-29738db6-7735-48b9-920a-e9d98d76aa24 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.024389] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129603, 'name': ReconfigVM_Task, 'duration_secs': 0.808525} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.024997] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Reconfigured VM instance instance-0000003a to attach disk [datastore1] d2a0e5d1-5557-41b6-936d-dc86d1346c61/d2a0e5d1-5557-41b6-936d-dc86d1346c61.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.025621] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83f57cb8-656c-425f-bf04-074eb5b635ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.032592] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 760.032592] env[61991]: value = "task-1129604" [ 760.032592] env[61991]: _type = "Task" [ 760.032592] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.041038] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129604, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.050742] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527965e6-6291-29cb-2351-cdfc85e49a08, 'name': SearchDatastore_Task, 'duration_secs': 0.014212} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.051105] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.051368] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.051616] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.051765] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.051956] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.052252] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa25b07e-0182-410a-93bc-915b14a4c792 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.062220] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.062449] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 760.063246] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68b54696-887c-4bbe-9035-4aea80f36dba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.068842] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 760.068842] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524ae661-021d-89b4-eae6-735491209008" [ 760.068842] env[61991]: _type = "Task" [ 760.068842] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.079093] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524ae661-021d-89b4-eae6-735491209008, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.149480] env[61991]: DEBUG nova.network.neutron [-] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.259193] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b262349-1212-4fd2-8cbf-991d114d9f24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.269537] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0477fdc-15e9-4620-82c9-b446602d02f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.282377] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.905s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.282933] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 760.287079] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.403s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.287079] env[61991]: DEBUG nova.objects.instance [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lazy-loading 'resources' on Instance uuid 30bd843a-9d86-4de1-b011-600e2a93bcb3 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.315818] env[61991]: DEBUG nova.compute.manager [req-9316df04-23db-4384-aed4-b9581a23b566 req-ac7741f4-97a2-4b95-90d5-4982cc9dcda1 service nova] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Detach interface failed, port_id=b1cd4427-17b1-40c2-8937-03d1f29e3d21, reason: Instance 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 760.329649] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.339161] env[61991]: DEBUG oslo_concurrency.lockutils [req-f90c692e-cb8e-4606-b189-4953b3bc446b req-04be60d5-b985-4de0-89f6-880ea70561a9 service nova] Releasing lock "refresh_cache-3e06e424-b64e-4fd2-8013-27760200c41d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.352895] env[61991]: DEBUG nova.network.neutron [-] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.389614] env[61991]: INFO nova.compute.manager [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Took 42.74 seconds to build instance. [ 760.401543] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 760.401788] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246849', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'name': 'volume-48ca7de5-1564-44be-83d4-3298210a3c5d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c17f844-1f0e-4f01-aadc-0f1f75a59d06', 'attached_at': '', 'detached_at': '', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'serial': '48ca7de5-1564-44be-83d4-3298210a3c5d'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 760.402963] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09428fdd-9976-4c06-b9d1-f1adb8843a0d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.421883] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957e6a1f-4cdc-4f90-81a5-517a47d38407 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.447404] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] volume-48ca7de5-1564-44be-83d4-3298210a3c5d/volume-48ca7de5-1564-44be-83d4-3298210a3c5d.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.449927] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13bef90f-66af-4663-9034-7b023c914c1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.469714] env[61991]: DEBUG oslo_vmware.api [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 760.469714] env[61991]: value = "task-1129605" [ 760.469714] env[61991]: _type = "Task" [ 760.469714] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.479052] env[61991]: DEBUG oslo_vmware.api [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.542685] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129604, 'name': Rename_Task, 'duration_secs': 0.311525} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.542985] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 760.543395] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-489d2e5b-a5c6-45fe-b6d4-b24a174a42fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.549630] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 760.549630] env[61991]: value = "task-1129606" [ 760.549630] env[61991]: _type = "Task" [ 760.549630] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.559230] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.580892] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524ae661-021d-89b4-eae6-735491209008, 'name': SearchDatastore_Task, 'duration_secs': 0.00975} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.581746] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f1c642-6f0b-4e6c-800e-a38d10cd1b0d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.587121] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 760.587121] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5247f983-2946-8947-fd0f-c192344e45fa" [ 760.587121] env[61991]: _type = "Task" [ 760.587121] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.595057] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5247f983-2946-8947-fd0f-c192344e45fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.653101] env[61991]: INFO nova.compute.manager [-] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Took 1.58 seconds to deallocate network for instance. [ 760.668138] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.670550] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fbbed0-75bf-4672-897c-30393cd4880a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.677724] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 760.678035] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26c9eabb-4026-4d0e-bcab-4bfbe3b282c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.746345] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 760.746345] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 760.746345] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleting the datastore file [datastore2] 78631e83-e574-4a90-b062-176b2d12d681 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.746749] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eafb8481-6892-4e81-8895-6ba90ca25beb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.758980] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 760.758980] env[61991]: value = "task-1129608" [ 760.758980] env[61991]: _type = "Task" [ 760.758980] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.769690] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.790800] env[61991]: DEBUG nova.compute.utils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.795139] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 760.797727] env[61991]: DEBUG nova.network.neutron [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 760.851679] env[61991]: DEBUG nova.policy [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5279cbbae361450fae512538226aa776', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9eb579be6d0d439db8465cc9639bdffb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.854044] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Successfully updated port: 35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 760.856310] env[61991]: INFO nova.compute.manager [-] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Took 1.53 seconds to deallocate network for instance. [ 760.893549] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec34e19b-57e9-4d34-b33b-787f40174842 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.330s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.983229] env[61991]: DEBUG oslo_vmware.api [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129605, 'name': ReconfigVM_Task, 'duration_secs': 0.408251} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.983325] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Reconfigured VM instance instance-0000002a to attach disk [datastore1] volume-48ca7de5-1564-44be-83d4-3298210a3c5d/volume-48ca7de5-1564-44be-83d4-3298210a3c5d.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.993285] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05522301-d10e-4496-95d0-61a3afa27bfb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.009973] env[61991]: DEBUG oslo_vmware.api [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 761.009973] env[61991]: value = "task-1129609" [ 761.009973] env[61991]: _type = "Task" [ 761.009973] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.018675] env[61991]: DEBUG oslo_vmware.api [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129609, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.061626] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129606, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.101487] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5247f983-2946-8947-fd0f-c192344e45fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009081} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.104690] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.105063] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997/7cbf26bd-51ab-41f7-976a-4ad9c3c1a997.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.105561] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16840163-0a8c-4c16-b0f7-802c2d3822e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.112418] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 761.112418] env[61991]: value = "task-1129610" [ 761.112418] env[61991]: _type = "Task" [ 761.112418] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.122652] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129610, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.165562] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.248069] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f9c45a-ba64-40bf-af3d-8899a1b31103 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.258898] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cf98b6-16c8-4350-a912-d38c1a2f1a2b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.276405] env[61991]: DEBUG oslo_vmware.api [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257359} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.306934] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.306934] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 761.306934] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.309141] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 761.314752] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27aa5cb5-eab6-4ff8-9c1c-41118e3c5a6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.315956] env[61991]: DEBUG nova.network.neutron [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Successfully created port: 179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.324581] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c05834-5453-417a-9476-eda7eba9884b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.344824] env[61991]: DEBUG nova.compute.provider_tree [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.344824] env[61991]: INFO nova.scheduler.client.report [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocations for instance 78631e83-e574-4a90-b062-176b2d12d681 [ 761.358630] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "refresh_cache-671e0197-4f96-4660-be99-ea1d1c0588e1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.359289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "refresh_cache-671e0197-4f96-4660-be99-ea1d1c0588e1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.359289] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.368662] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.396610] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.519624] env[61991]: DEBUG oslo_vmware.api [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129609, 'name': ReconfigVM_Task, 'duration_secs': 0.19217} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.520180] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246849', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'name': 'volume-48ca7de5-1564-44be-83d4-3298210a3c5d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c17f844-1f0e-4f01-aadc-0f1f75a59d06', 'attached_at': '', 'detached_at': '', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'serial': '48ca7de5-1564-44be-83d4-3298210a3c5d'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 761.561655] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129606, 'name': PowerOnVM_Task, 'duration_secs': 0.648259} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.561936] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 761.562193] env[61991]: INFO nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Took 9.34 seconds to spawn the instance on the hypervisor. [ 761.562411] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 761.565642] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee11ba6-cf09-48f0-ab12-c9abdeb3566c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.623362] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129610, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.853283] env[61991]: DEBUG nova.scheduler.client.report [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.856663] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.859080] env[61991]: DEBUG nova.compute.manager [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Received event network-vif-plugged-35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.859773] env[61991]: DEBUG oslo_concurrency.lockutils [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] Acquiring lock "671e0197-4f96-4660-be99-ea1d1c0588e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.860079] env[61991]: DEBUG oslo_concurrency.lockutils [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.860318] env[61991]: DEBUG oslo_concurrency.lockutils [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.860529] env[61991]: DEBUG nova.compute.manager [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] No waiting events found dispatching network-vif-plugged-35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 761.861503] env[61991]: WARNING nova.compute.manager [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Received unexpected event network-vif-plugged-35afc727-b479-4f5e-b793-9e8be36b9975 for instance with vm_state building and task_state spawning. [ 761.861746] env[61991]: DEBUG nova.compute.manager [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Received event network-changed-35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.861949] env[61991]: DEBUG nova.compute.manager [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Refreshing instance network info cache due to event network-changed-35afc727-b479-4f5e-b793-9e8be36b9975. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.862264] env[61991]: DEBUG oslo_concurrency.lockutils [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] Acquiring lock "refresh_cache-671e0197-4f96-4660-be99-ea1d1c0588e1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.925024] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.062515] env[61991]: DEBUG nova.compute.manager [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Received event network-vif-unplugged-8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.062757] env[61991]: DEBUG oslo_concurrency.lockutils [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] Acquiring lock "78631e83-e574-4a90-b062-176b2d12d681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.063018] env[61991]: DEBUG oslo_concurrency.lockutils [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] Lock "78631e83-e574-4a90-b062-176b2d12d681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.063261] env[61991]: DEBUG oslo_concurrency.lockutils [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] Lock "78631e83-e574-4a90-b062-176b2d12d681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.063379] env[61991]: DEBUG nova.compute.manager [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] No waiting events found dispatching network-vif-unplugged-8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 762.063551] env[61991]: WARNING nova.compute.manager [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Received unexpected event network-vif-unplugged-8aa50c96-062e-4d6c-9aa0-990a50838237 for instance with vm_state shelved_offloaded and task_state None. [ 762.063713] env[61991]: DEBUG nova.compute.manager [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Received event network-changed-8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.063866] env[61991]: DEBUG nova.compute.manager [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Refreshing instance network info cache due to event network-changed-8aa50c96-062e-4d6c-9aa0-990a50838237. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.064871] env[61991]: DEBUG oslo_concurrency.lockutils [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] Acquiring lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.065055] env[61991]: DEBUG oslo_concurrency.lockutils [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] Acquired lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.065229] env[61991]: DEBUG nova.network.neutron [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Refreshing network info cache for port 8aa50c96-062e-4d6c-9aa0-990a50838237 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 762.089081] env[61991]: INFO nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Took 35.38 seconds to build instance. [ 762.124548] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129610, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.906508} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.124848] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997/7cbf26bd-51ab-41f7-976a-4ad9c3c1a997.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 762.125076] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.125413] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65b86132-8cd2-4c65-aa73-752c2b42f53e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.128321] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.134981] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 762.134981] env[61991]: value = "task-1129611" [ 762.134981] env[61991]: _type = "Task" [ 762.134981] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.145920] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.207885] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "7df78da9-bf65-4621-b50d-43f1d721c2f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.208181] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.208714] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "7df78da9-bf65-4621-b50d-43f1d721c2f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.208714] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.208838] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.211290] env[61991]: INFO nova.compute.manager [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Terminating instance [ 762.215423] env[61991]: DEBUG nova.compute.manager [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.215647] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.216778] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12cb74b-ccd4-4703-834a-23c4399fd47d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.226115] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.226115] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c4a6231-d942-4475-a7b4-41ed40e952d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.233716] env[61991]: DEBUG oslo_vmware.api [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 762.233716] env[61991]: value = "task-1129612" [ 762.233716] env[61991]: _type = "Task" [ 762.233716] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.245867] env[61991]: DEBUG oslo_vmware.api [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129612, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.326809] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.353782] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.354043] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.354200] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.354417] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.354573] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.354721] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.354928] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.355609] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.355609] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.355785] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.357106] env[61991]: DEBUG nova.virt.hardware [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.357106] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c172c7d-d69c-45f2-91fe-73d17c2c75f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.359846] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.073s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.361911] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.662s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.367905] env[61991]: INFO nova.compute.claims [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.382432] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b5f9ac-71da-475a-9526-4c07b9696fc6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.391577] env[61991]: INFO nova.scheduler.client.report [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Deleted allocations for instance 30bd843a-9d86-4de1-b011-600e2a93bcb3 [ 762.432081] env[61991]: DEBUG nova.network.neutron [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Updating instance_info_cache with network_info: [{"id": "35afc727-b479-4f5e-b793-9e8be36b9975", "address": "fa:16:3e:7b:7e:8c", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35afc727-b4", "ovs_interfaceid": "35afc727-b479-4f5e-b793-9e8be36b9975", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.589733] env[61991]: DEBUG nova.objects.instance [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'flavor' on Instance uuid 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.593993] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.907s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.647966] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068499} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.648296] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.649287] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b99f4b-7bd0-41ee-a3e3-6325e70b8739 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.676174] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997/7cbf26bd-51ab-41f7-976a-4ad9c3c1a997.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.676979] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8461eda-906d-4f5c-bd2d-20286879b718 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.698845] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 762.698845] env[61991]: value = "task-1129613" [ 762.698845] env[61991]: _type = "Task" [ 762.698845] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.711161] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.743614] env[61991]: DEBUG oslo_vmware.api [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129612, 'name': PowerOffVM_Task, 'duration_secs': 0.202433} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.744833] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.744833] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.745751] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4a0b675-1e2d-4098-b95e-b2619b071492 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.809386] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.809634] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.809887] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Deleting the datastore file [datastore1] 7df78da9-bf65-4621-b50d-43f1d721c2f1 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.810328] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-375e5a4a-ca78-443d-a8a8-582e8ab72fbb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.818615] env[61991]: DEBUG oslo_vmware.api [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for the task: (returnval){ [ 762.818615] env[61991]: value = "task-1129615" [ 762.818615] env[61991]: _type = "Task" [ 762.818615] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.850039] env[61991]: DEBUG oslo_vmware.api [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129615, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.915325] env[61991]: DEBUG oslo_concurrency.lockutils [None req-91b173a2-8036-4c0c-b4fe-91f95ecf5d61 tempest-ServerGroupTestJSON-885351965 tempest-ServerGroupTestJSON-885351965-project-member] Lock "30bd843a-9d86-4de1-b011-600e2a93bcb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.552s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.934712] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "refresh_cache-671e0197-4f96-4660-be99-ea1d1c0588e1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.935097] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Instance network_info: |[{"id": "35afc727-b479-4f5e-b793-9e8be36b9975", "address": "fa:16:3e:7b:7e:8c", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35afc727-b4", "ovs_interfaceid": "35afc727-b479-4f5e-b793-9e8be36b9975", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 762.935439] env[61991]: DEBUG oslo_concurrency.lockutils [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] Acquired lock "refresh_cache-671e0197-4f96-4660-be99-ea1d1c0588e1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.935993] env[61991]: DEBUG nova.network.neutron [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Refreshing network info cache for port 35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 762.937608] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:7e:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35afc727-b479-4f5e-b793-9e8be36b9975', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.950347] env[61991]: DEBUG oslo.service.loopingcall [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.956549] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 762.956549] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20c401eb-38bc-48b4-a6ea-c8452599cd35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.985355] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.985355] env[61991]: value = "task-1129616" [ 762.985355] env[61991]: _type = "Task" [ 762.985355] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.996673] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129616, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.076841] env[61991]: DEBUG nova.network.neutron [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Updated VIF entry in instance network info cache for port 8aa50c96-062e-4d6c-9aa0-990a50838237. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 763.077273] env[61991]: DEBUG nova.network.neutron [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Updating instance_info_cache with network_info: [{"id": "8aa50c96-062e-4d6c-9aa0-990a50838237", "address": "fa:16:3e:62:a0:f9", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": null, "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8aa50c96-06", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.097992] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.100624] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6b611ce7-03d3-45d2-83a9-0b28295a0105 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.370s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.214674] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129613, 'name': ReconfigVM_Task, 'duration_secs': 0.331723} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.214674] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997/7cbf26bd-51ab-41f7-976a-4ad9c3c1a997.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.214674] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6b1200a-7b3b-4e11-ad98-7f0d9a3da138 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.220298] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 763.220298] env[61991]: value = "task-1129617" [ 763.220298] env[61991]: _type = "Task" [ 763.220298] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.230453] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129617, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.329273] env[61991]: DEBUG oslo_vmware.api [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Task: {'id': task-1129615, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205416} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.329273] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.329273] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.329273] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.329273] env[61991]: INFO nova.compute.manager [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 763.329273] env[61991]: DEBUG oslo.service.loopingcall [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.329273] env[61991]: DEBUG nova.compute.manager [-] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.329666] env[61991]: DEBUG nova.network.neutron [-] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.340705] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "78631e83-e574-4a90-b062-176b2d12d681" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.499332] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129616, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.581869] env[61991]: DEBUG oslo_concurrency.lockutils [req-2c44cc87-3812-4611-804b-9ac22c1b33e9 req-7254beb9-6e32-4800-b0a2-90510947b5fd service nova] Releasing lock "refresh_cache-78631e83-e574-4a90-b062-176b2d12d681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.623566] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.735093] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129617, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.753160] env[61991]: DEBUG nova.network.neutron [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Successfully updated port: 179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 763.807672] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80518a45-9a5f-44a0-98b8-ba9d601de986 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.817047] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f49e3f-6972-4602-b847-13e572e826f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.851090] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8c7fd7-3599-4619-8927-025743b5794b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.862292] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b1c521-4b57-43ab-9dca-48c4bf5d0ce9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.881089] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.994898] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129616, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.031570] env[61991]: DEBUG nova.compute.manager [req-ebae101d-42f5-4765-a20f-faf842ec2d7e req-438d8627-8133-4c6e-8b73-8342cf5156cb service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Received event network-vif-deleted-248dc7ce-d07a-4030-810d-e49b969cb8ab {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.031779] env[61991]: INFO nova.compute.manager [req-ebae101d-42f5-4765-a20f-faf842ec2d7e req-438d8627-8133-4c6e-8b73-8342cf5156cb service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Neutron deleted interface 248dc7ce-d07a-4030-810d-e49b969cb8ab; detaching it from the instance and deleting it from the info cache [ 764.032080] env[61991]: DEBUG nova.network.neutron [req-ebae101d-42f5-4765-a20f-faf842ec2d7e req-438d8627-8133-4c6e-8b73-8342cf5156cb service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.061028] env[61991]: DEBUG nova.network.neutron [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Updated VIF entry in instance network info cache for port 35afc727-b479-4f5e-b793-9e8be36b9975. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 764.061433] env[61991]: DEBUG nova.network.neutron [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Updating instance_info_cache with network_info: [{"id": "35afc727-b479-4f5e-b793-9e8be36b9975", "address": "fa:16:3e:7b:7e:8c", "network": {"id": "b5f7224c-ed2e-415d-8468-36d669f405b6", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-509722275-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a44a2e26d9f45bcba0cdeecf09a04c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35afc727-b4", "ovs_interfaceid": "35afc727-b479-4f5e-b793-9e8be36b9975", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.230901] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129617, 'name': Rename_Task, 'duration_secs': 0.894545} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.231377] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 764.231760] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f60460f9-ca6f-442e-8745-fb969fce5421 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.240047] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 764.240047] env[61991]: value = "task-1129618" [ 764.240047] env[61991]: _type = "Task" [ 764.240047] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.246142] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.255943] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.256339] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquired lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.256577] env[61991]: DEBUG nova.network.neutron [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.268744] env[61991]: DEBUG nova.compute.manager [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Received event network-vif-plugged-179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.268991] env[61991]: DEBUG oslo_concurrency.lockutils [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] Acquiring lock "d6a2d226-5097-4461-b2ce-fa698f1066d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.269651] env[61991]: DEBUG oslo_concurrency.lockutils [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.269856] env[61991]: DEBUG oslo_concurrency.lockutils [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.270405] env[61991]: DEBUG nova.compute.manager [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] No waiting events found dispatching network-vif-plugged-179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 764.270692] env[61991]: WARNING nova.compute.manager [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Received unexpected event network-vif-plugged-179a25d5-af58-45b5-b510-6f9064bddb3c for instance with vm_state building and task_state spawning. [ 764.270955] env[61991]: DEBUG nova.compute.manager [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Received event network-changed-179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.271186] env[61991]: DEBUG nova.compute.manager [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Refreshing instance network info cache due to event network-changed-179a25d5-af58-45b5-b510-6f9064bddb3c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 764.272501] env[61991]: DEBUG oslo_concurrency.lockutils [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] Acquiring lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.386943] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.465910] env[61991]: DEBUG nova.network.neutron [-] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.494439] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129616, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.534706] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dce556b0-9254-47b2-9d98-545b5dbf1040 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.549931] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2873e936-492b-45e9-89c1-ca7a1f248b6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.567049] env[61991]: DEBUG oslo_concurrency.lockutils [req-1bba1e56-9084-4d04-88f7-efa6d4978555 req-082ff8e3-16fa-4dd1-b54e-a99279d55533 service nova] Releasing lock "refresh_cache-671e0197-4f96-4660-be99-ea1d1c0588e1" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.588877] env[61991]: DEBUG nova.compute.manager [req-ebae101d-42f5-4765-a20f-faf842ec2d7e req-438d8627-8133-4c6e-8b73-8342cf5156cb service nova] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Detach interface failed, port_id=248dc7ce-d07a-4030-810d-e49b969cb8ab, reason: Instance 7df78da9-bf65-4621-b50d-43f1d721c2f1 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 764.753597] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129618, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.811203] env[61991]: DEBUG nova.network.neutron [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.876388] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.876858] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.877377] env[61991]: INFO nova.compute.manager [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Shelving [ 764.893501] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.894372] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.896877] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.240s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.899824] env[61991]: INFO nova.compute.claims [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.970818] env[61991]: INFO nova.compute.manager [-] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Took 1.64 seconds to deallocate network for instance. [ 764.997564] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129616, 'name': CreateVM_Task, 'duration_secs': 1.670373} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.997753] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 764.998444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.998618] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.998939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 764.999220] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-454af177-afcd-4956-aade-4d4312a9d357 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.004292] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 765.004292] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528d15aa-f61d-5cd8-035b-6c13764d0e6b" [ 765.004292] env[61991]: _type = "Task" [ 765.004292] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.014050] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528d15aa-f61d-5cd8-035b-6c13764d0e6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.100847] env[61991]: DEBUG nova.network.neutron [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updating instance_info_cache with network_info: [{"id": "179a25d5-af58-45b5-b510-6f9064bddb3c", "address": "fa:16:3e:30:74:a2", "network": {"id": "4dff2d72-8dde-4106-bd50-da1177efb4bd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1324618604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9eb579be6d0d439db8465cc9639bdffb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap179a25d5-af", "ovs_interfaceid": "179a25d5-af58-45b5-b510-6f9064bddb3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.256493] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129618, 'name': PowerOnVM_Task, 'duration_secs': 0.612364} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.256493] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 765.256493] env[61991]: INFO nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Took 10.49 seconds to spawn the instance on the hypervisor. [ 765.256493] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 765.256619] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be2d8ba-bd17-42c5-b9dc-ad2e5e13acbd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.390422] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 765.390762] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f698b7c-c515-42cc-942d-6d15badd6c8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.399718] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 765.399718] env[61991]: value = "task-1129619" [ 765.399718] env[61991]: _type = "Task" [ 765.399718] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.406542] env[61991]: DEBUG nova.compute.utils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.415747] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.415747] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.418628] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.478681] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.504406] env[61991]: DEBUG nova.policy [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c54258ac7b3e493d808c496531ca4d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fd3b8e07dce44a8b03543c3c0ca1bb5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.517541] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528d15aa-f61d-5cd8-035b-6c13764d0e6b, 'name': SearchDatastore_Task, 'duration_secs': 0.009171} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.517896] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.518210] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.518505] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.518749] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.518963] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.519334] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-266c933f-c261-418f-a325-b379f1a2e2e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.528298] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.528505] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 765.529286] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-440e0abe-b1e7-49e0-a5d1-0ab622a6bae1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.536939] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 765.536939] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52608a2a-828a-9777-4677-1040c947f971" [ 765.536939] env[61991]: _type = "Task" [ 765.536939] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.547667] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.548033] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.553271] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52608a2a-828a-9777-4677-1040c947f971, 'name': SearchDatastore_Task, 'duration_secs': 0.009093} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.554523] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4095b816-202f-4f95-ba40-00e7cd027ee5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.561027] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 765.561027] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52862c5d-a6ce-6439-57d6-9bfdff85c952" [ 765.561027] env[61991]: _type = "Task" [ 765.561027] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.577909] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52862c5d-a6ce-6439-57d6-9bfdff85c952, 'name': SearchDatastore_Task, 'duration_secs': 0.009687} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.577909] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.578153] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 671e0197-4f96-4660-be99-ea1d1c0588e1/671e0197-4f96-4660-be99-ea1d1c0588e1.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 765.579415] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1bd21297-8a35-4806-b1b3-56e495b16fa0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.586193] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 765.586193] env[61991]: value = "task-1129620" [ 765.586193] env[61991]: _type = "Task" [ 765.586193] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.594709] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.603547] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Releasing lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.604490] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Instance network_info: |[{"id": "179a25d5-af58-45b5-b510-6f9064bddb3c", "address": "fa:16:3e:30:74:a2", "network": {"id": "4dff2d72-8dde-4106-bd50-da1177efb4bd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1324618604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9eb579be6d0d439db8465cc9639bdffb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap179a25d5-af", "ovs_interfaceid": "179a25d5-af58-45b5-b510-6f9064bddb3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 765.604490] env[61991]: DEBUG oslo_concurrency.lockutils [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] Acquired lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.604490] env[61991]: DEBUG nova.network.neutron [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Refreshing network info cache for port 179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.606075] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:74:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '179a25d5-af58-45b5-b510-6f9064bddb3c', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.615036] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Creating folder: Project (9eb579be6d0d439db8465cc9639bdffb). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.615648] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-094e072c-38cf-4506-af4a-45361ea4bafb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.627725] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Created folder: Project (9eb579be6d0d439db8465cc9639bdffb) in parent group-v246753. [ 765.627725] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Creating folder: Instances. Parent ref: group-v246852. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.627725] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5be6c260-6275-444c-8160-e39f20109780 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.636875] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Created folder: Instances in parent group-v246852. [ 765.637174] env[61991]: DEBUG oslo.service.loopingcall [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.637388] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 765.637614] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df727226-794f-4a51-8051-382e5982f857 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.661361] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.661361] env[61991]: value = "task-1129623" [ 765.661361] env[61991]: _type = "Task" [ 765.661361] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.669952] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129623, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.780061] env[61991]: INFO nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Took 38.99 seconds to build instance. [ 765.912701] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129619, 'name': PowerOffVM_Task, 'duration_secs': 0.287188} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.913216] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.914117] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33299edf-c998-49e0-af4c-83fcb7db9e2c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.918022] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.947711] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8b7dbf-f545-4dcb-8893-a9ed42bf39b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.064855] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Successfully created port: 6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.097242] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129620, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493335} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.097794] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 671e0197-4f96-4660-be99-ea1d1c0588e1/671e0197-4f96-4660-be99-ea1d1c0588e1.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 766.097794] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.097969] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf7f9f25-ae87-4bc0-a912-6e333660fb48 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.105100] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 766.105100] env[61991]: value = "task-1129624" [ 766.105100] env[61991]: _type = "Task" [ 766.105100] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.121521] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.172734] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129623, 'name': CreateVM_Task, 'duration_secs': 0.463452} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.175817] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 766.175817] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.175817] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.175817] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 766.175817] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cc87bd7-541c-408c-874b-b0301537d2fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.179918] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 766.179918] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d84510-016e-e22d-bef2-3bf0076c1a4b" [ 766.179918] env[61991]: _type = "Task" [ 766.179918] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.196859] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d84510-016e-e22d-bef2-3bf0076c1a4b, 'name': SearchDatastore_Task, 'duration_secs': 0.009412} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.197373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.197710] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.198228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.198342] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.198744] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.198957] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a468ae01-cfe0-43ca-8f30-d7d8ffe4f3d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.210550] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.210550] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 766.210741] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80499562-46e5-463f-80a2-4ce70c995e85 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.216779] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 766.216779] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520f0d28-35f2-0565-7455-2adb9da21cdb" [ 766.216779] env[61991]: _type = "Task" [ 766.216779] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.223873] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520f0d28-35f2-0565-7455-2adb9da21cdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.285851] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.571s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.395482] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889b48d5-8f0f-4a01-9723-237c9df185d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.403666] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c173f7-e608-420a-ad5c-54985fad42b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.443591] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a5c19e-71fe-4f80-a37d-b38f825405d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.451970] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0063f563-a003-4451-9b19-a95f2fd76b65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.460400] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 766.470492] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-453f15b0-c6fb-4a17-b55a-66c0ab316610 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.475229] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 766.482676] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 766.482676] env[61991]: value = "task-1129625" [ 766.482676] env[61991]: _type = "Task" [ 766.482676] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.493566] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129625, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.618358] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066095} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.619270] env[61991]: DEBUG nova.network.neutron [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updated VIF entry in instance network info cache for port 179a25d5-af58-45b5-b510-6f9064bddb3c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 766.619607] env[61991]: DEBUG nova.network.neutron [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updating instance_info_cache with network_info: [{"id": "179a25d5-af58-45b5-b510-6f9064bddb3c", "address": "fa:16:3e:30:74:a2", "network": {"id": "4dff2d72-8dde-4106-bd50-da1177efb4bd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1324618604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9eb579be6d0d439db8465cc9639bdffb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap179a25d5-af", "ovs_interfaceid": "179a25d5-af58-45b5-b510-6f9064bddb3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.620812] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.621835] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2885c87-be42-4b29-bd84-ecee4ca1bda5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.647838] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 671e0197-4f96-4660-be99-ea1d1c0588e1/671e0197-4f96-4660-be99-ea1d1c0588e1.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.648808] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69f68c0b-dafc-4d54-8b6d-5543a0610695 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.671593] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 766.671593] env[61991]: value = "task-1129626" [ 766.671593] env[61991]: _type = "Task" [ 766.671593] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.685260] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129626, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.726081] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520f0d28-35f2-0565-7455-2adb9da21cdb, 'name': SearchDatastore_Task, 'duration_secs': 0.008718} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.726941] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cba374b-57a2-4246-a0a4-025da824ec0a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.731863] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 766.731863] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d07c2e-99b7-a8b9-da05-3899f736b158" [ 766.731863] env[61991]: _type = "Task" [ 766.731863] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.739681] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d07c2e-99b7-a8b9-da05-3899f736b158, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.794606] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.945014] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.972527] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.972833] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.972998] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.973205] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.973358] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.973509] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.973720] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.973880] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.974075] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.974250] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.977237] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.977237] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f3e75e-05f3-4afc-b201-d6079aef4cff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.988715] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da184726-92fb-4559-abe9-9ced266ad25f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.997811] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129625, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.007255] env[61991]: ERROR nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [req-64451440-cd02-4f76-bd41-935aa06bb9a8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-64451440-cd02-4f76-bd41-935aa06bb9a8"}]} [ 767.029066] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 767.042417] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 767.042593] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 767.057246] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 767.081758] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 767.123915] env[61991]: DEBUG oslo_concurrency.lockutils [req-5501b323-c240-42bd-b331-6afe18b12bc3 req-6a7df702-8742-42a3-960b-68f74a3bf5a9 service nova] Releasing lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.184328] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129626, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.246685] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d07c2e-99b7-a8b9-da05-3899f736b158, 'name': SearchDatastore_Task, 'duration_secs': 0.008976} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.249381] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.249658] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] d6a2d226-5097-4461-b2ce-fa698f1066d5/d6a2d226-5097-4461-b2ce-fa698f1066d5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.250119] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1fb6249-5ba6-4b90-9d32-7d66a43da40c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.257546] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 767.257546] env[61991]: value = "task-1129627" [ 767.257546] env[61991]: _type = "Task" [ 767.257546] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.268553] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129627, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.318302] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.453815] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f003b1e-78e7-47ba-820b-5440a5273785 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.465245] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c642cd-ee58-48ac-a6fb-4b205b9ba762 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.508371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6225d752-758e-46ac-a6ac-4063e69b6d9d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.522250] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129625, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.524568] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db145437-df3b-4010-8fdb-3a9d653fa49a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.547291] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 767.628663] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.628935] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.684860] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129626, 'name': ReconfigVM_Task, 'duration_secs': 0.798228} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.685105] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 671e0197-4f96-4660-be99-ea1d1c0588e1/671e0197-4f96-4660-be99-ea1d1c0588e1.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.685847] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc4753a2-62a1-40be-bf4f-950a6fcfa00d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.692956] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 767.692956] env[61991]: value = "task-1129628" [ 767.692956] env[61991]: _type = "Task" [ 767.692956] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.702730] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129628, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.767777] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129627, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.809632] env[61991]: DEBUG nova.compute.manager [req-68f8ce7f-343e-40d8-bd5e-c08e1d6fe3c1 req-58b8f16b-d118-4398-a9da-f8aabdc34c48 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Received event network-vif-plugged-6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.810035] env[61991]: DEBUG oslo_concurrency.lockutils [req-68f8ce7f-343e-40d8-bd5e-c08e1d6fe3c1 req-58b8f16b-d118-4398-a9da-f8aabdc34c48 service nova] Acquiring lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.810224] env[61991]: DEBUG oslo_concurrency.lockutils [req-68f8ce7f-343e-40d8-bd5e-c08e1d6fe3c1 req-58b8f16b-d118-4398-a9da-f8aabdc34c48 service nova] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.810512] env[61991]: DEBUG oslo_concurrency.lockutils [req-68f8ce7f-343e-40d8-bd5e-c08e1d6fe3c1 req-58b8f16b-d118-4398-a9da-f8aabdc34c48 service nova] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.810758] env[61991]: DEBUG nova.compute.manager [req-68f8ce7f-343e-40d8-bd5e-c08e1d6fe3c1 req-58b8f16b-d118-4398-a9da-f8aabdc34c48 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] No waiting events found dispatching network-vif-plugged-6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 767.810994] env[61991]: WARNING nova.compute.manager [req-68f8ce7f-343e-40d8-bd5e-c08e1d6fe3c1 req-58b8f16b-d118-4398-a9da-f8aabdc34c48 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Received unexpected event network-vif-plugged-6a28231d-a28f-49ea-8339-3ca4a02684bb for instance with vm_state building and task_state spawning. [ 767.909024] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Successfully updated port: 6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 768.013188] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129625, 'name': CreateSnapshot_Task, 'duration_secs': 1.047504} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.013515] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 768.014312] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52284202-d98c-4033-a41c-bf1d06582bd8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.070206] env[61991]: ERROR nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [req-b36234f8-05e3-4d09-85dc-ba8068a4c0bb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b36234f8-05e3-4d09-85dc-ba8068a4c0bb"}]} [ 768.088404] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 768.105289] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 768.105416] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.122128] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 768.142268] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 768.211554] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129628, 'name': Rename_Task, 'duration_secs': 0.39986} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.212733] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 768.214020] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3276e5b6-ed97-4d8b-843d-897c7a0f2c64 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.222525] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 768.222525] env[61991]: value = "task-1129629" [ 768.222525] env[61991]: _type = "Task" [ 768.222525] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.230380] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.270068] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129627, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549156} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.272608] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] d6a2d226-5097-4461-b2ce-fa698f1066d5/d6a2d226-5097-4461-b2ce-fa698f1066d5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 768.272838] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.273347] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0b0e8e3-7eb6-419a-96c3-8f8e5fc4af47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.279275] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 768.279275] env[61991]: value = "task-1129630" [ 768.279275] env[61991]: _type = "Task" [ 768.279275] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.289249] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129630, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.412498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "refresh_cache-0aa67bb1-59d3-42e4-8f12-47e61a130dfc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.412498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "refresh_cache-0aa67bb1-59d3-42e4-8f12-47e61a130dfc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.412498] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.495513] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d984c5d-380c-4ce2-94f1-d00d418247ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.503202] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925a8c76-559c-4a14-a98c-dceb71c4b6aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.540757] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 768.541315] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-74308f26-9303-4260-b50a-2ed1f602d2fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.544866] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa7c603-57f0-4b74-b563-60659960a9f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.554678] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0f9d7e-96b1-4ba4-b5a2-69c1c874dc88 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.558742] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 768.558742] env[61991]: value = "task-1129631" [ 768.558742] env[61991]: _type = "Task" [ 768.558742] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.570026] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.577672] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129631, 'name': CloneVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.733157] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129629, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.792939] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129630, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066632} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.793118] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.793879] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4204665-a14c-4690-8fd8-89d6537af2b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.815754] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] d6a2d226-5097-4461-b2ce-fa698f1066d5/d6a2d226-5097-4461-b2ce-fa698f1066d5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.816377] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b63269f-015b-48b5-ae11-e5f752e23279 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.836114] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 768.836114] env[61991]: value = "task-1129632" [ 768.836114] env[61991]: _type = "Task" [ 768.836114] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.845014] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129632, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.950559] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.071244] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129631, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.103309] env[61991]: DEBUG nova.scheduler.client.report [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 80 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 769.103616] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 80 to 81 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 769.103855] env[61991]: DEBUG nova.compute.provider_tree [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 769.170152] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Updating instance_info_cache with network_info: [{"id": "6a28231d-a28f-49ea-8339-3ca4a02684bb", "address": "fa:16:3e:15:a3:26", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a28231d-a2", "ovs_interfaceid": "6a28231d-a28f-49ea-8339-3ca4a02684bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.233717] env[61991]: DEBUG oslo_vmware.api [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129629, 'name': PowerOnVM_Task, 'duration_secs': 0.784215} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.234900] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 769.234900] env[61991]: INFO nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Took 9.81 seconds to spawn the instance on the hypervisor. [ 769.234900] env[61991]: DEBUG nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 769.235642] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ff1b28-5c9c-418c-b0a9-710faf2a864a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.347016] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129632, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.573011] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129631, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.610107] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.713s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.610727] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.613410] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 29.093s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.674757] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "refresh_cache-0aa67bb1-59d3-42e4-8f12-47e61a130dfc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.675906] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Instance network_info: |[{"id": "6a28231d-a28f-49ea-8339-3ca4a02684bb", "address": "fa:16:3e:15:a3:26", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a28231d-a2", "ovs_interfaceid": "6a28231d-a28f-49ea-8339-3ca4a02684bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 769.675906] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:a3:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a28231d-a28f-49ea-8339-3ca4a02684bb', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.685429] env[61991]: DEBUG oslo.service.loopingcall [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.687756] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.688247] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca96ee2e-151e-41a9-9333-1bd2ca7b18e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.710318] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.710318] env[61991]: value = "task-1129633" [ 769.710318] env[61991]: _type = "Task" [ 769.710318] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.719130] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129633, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.760034] env[61991]: INFO nova.compute.manager [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Took 38.89 seconds to build instance. [ 769.849578] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129632, 'name': ReconfigVM_Task, 'duration_secs': 0.66171} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.850052] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Reconfigured VM instance instance-0000003d to attach disk [datastore2] d6a2d226-5097-4461-b2ce-fa698f1066d5/d6a2d226-5097-4461-b2ce-fa698f1066d5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.850864] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7145950c-ec5b-4c23-b9d9-29122017b8cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.858023] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 769.858023] env[61991]: value = "task-1129634" [ 769.858023] env[61991]: _type = "Task" [ 769.858023] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.867331] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129634, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.947615] env[61991]: DEBUG nova.compute.manager [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Received event network-changed-6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.947928] env[61991]: DEBUG nova.compute.manager [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Refreshing instance network info cache due to event network-changed-6a28231d-a28f-49ea-8339-3ca4a02684bb. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.948248] env[61991]: DEBUG oslo_concurrency.lockutils [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] Acquiring lock "refresh_cache-0aa67bb1-59d3-42e4-8f12-47e61a130dfc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.948478] env[61991]: DEBUG oslo_concurrency.lockutils [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] Acquired lock "refresh_cache-0aa67bb1-59d3-42e4-8f12-47e61a130dfc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.948725] env[61991]: DEBUG nova.network.neutron [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Refreshing network info cache for port 6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 770.073968] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129631, 'name': CloneVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.127107] env[61991]: DEBUG nova.compute.utils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.128538] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.128706] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 770.212572] env[61991]: DEBUG nova.policy [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c54258ac7b3e493d808c496531ca4d5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fd3b8e07dce44a8b03543c3c0ca1bb5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.223294] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129633, 'name': CreateVM_Task, 'duration_secs': 0.314248} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.223462] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 770.224122] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.224300] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.224668] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 770.224922] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b93d019-dac5-4b12-9153-5ffe3dc930e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.229868] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 770.229868] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d67984-c2ca-51e3-2194-674c04468257" [ 770.229868] env[61991]: _type = "Task" [ 770.229868] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.237918] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d67984-c2ca-51e3-2194-674c04468257, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.261574] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fb34f886-5c50-4eee-b9bb-ef82fa4b6727 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.515s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.272183] env[61991]: DEBUG nova.compute.manager [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.273094] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260c1744-a88b-4d65-bf0d-75a40ab12b66 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.369040] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129634, 'name': Rename_Task, 'duration_secs': 0.185148} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.369359] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 770.369611] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75bb3582-c5d4-4553-9704-ca75f4a6663a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.376179] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 770.376179] env[61991]: value = "task-1129635" [ 770.376179] env[61991]: _type = "Task" [ 770.376179] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.384539] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.580978] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129631, 'name': CloneVM_Task, 'duration_secs': 1.620732} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.581453] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Created linked-clone VM from snapshot [ 770.582524] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38916c55-401b-419c-872a-ed3bb618686f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.592573] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Uploading image 0810354d-ca0b-4776-8487-42797b713043 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 770.617763] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 770.617763] env[61991]: value = "vm-246856" [ 770.617763] env[61991]: _type = "VirtualMachine" [ 770.617763] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 770.618429] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-27bfe8d2-b9eb-4ebe-829a-a082b154fddb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.632394] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Applying migration context for instance c14d99a3-950d-44d0-b330-3031a1f8a2be as it has an incoming, in-progress migration 528ff66f-4caf-4189-ac3a-ace614777b67. Migration status is reverting {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 770.634232] env[61991]: INFO nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating resource usage from migration 528ff66f-4caf-4189-ac3a-ace614777b67 [ 770.641624] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lease: (returnval){ [ 770.641624] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52305cb5-f943-4a2c-94f6-4b161bce0680" [ 770.641624] env[61991]: _type = "HttpNfcLease" [ 770.641624] env[61991]: } obtained for exporting VM: (result){ [ 770.641624] env[61991]: value = "vm-246856" [ 770.641624] env[61991]: _type = "VirtualMachine" [ 770.641624] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 770.641938] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the lease: (returnval){ [ 770.641938] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52305cb5-f943-4a2c-94f6-4b161bce0680" [ 770.641938] env[61991]: _type = "HttpNfcLease" [ 770.641938] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 770.643388] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.657389] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 770.657389] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52305cb5-f943-4a2c-94f6-4b161bce0680" [ 770.657389] env[61991]: _type = "HttpNfcLease" [ 770.657389] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 770.657972] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 770.657972] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52305cb5-f943-4a2c-94f6-4b161bce0680" [ 770.657972] env[61991]: _type = "HttpNfcLease" [ 770.657972] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 770.658679] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ce5e8a-b83a-43f9-b87a-c501cff2ff95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.667446] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Successfully created port: 2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.674706] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance acef131a-4263-42ad-964e-bb1e0cb21eec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 770.674706] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.674706] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 1715711a-72c5-4e86-88e0-cecf922cb42f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.674706] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.674892] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f78ef63b-453e-45d3-959b-4b0c1922b53e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.674892] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3bcde671-5702-4b8a-8881-88eb7dfd0556 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.674988] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 80d05278-b516-4408-94b0-11bc93500b5c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.675115] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 700ce4a7-d597-449a-9379-0cfb0c8f82fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.675243] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3e06e424-b64e-4fd2-8013-27760200c41d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 770.675364] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 770.675487] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 7df78da9-bf65-4621-b50d-43f1d721c2f1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 770.675600] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d2a0e5d1-5557-41b6-936d-dc86d1346c61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.675711] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.675819] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 671e0197-4f96-4660-be99-ea1d1c0588e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.676014] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d6a2d226-5097-4461-b2ce-fa698f1066d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.676072] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 0aa67bb1-59d3-42e4-8f12-47e61a130dfc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.676162] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c279a82f-bf4c-4a59-b04a-38d31070b0e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.676265] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Migration 528ff66f-4caf-4189-ac3a-ace614777b67 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 770.676374] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c14d99a3-950d-44d0-b330-3031a1f8a2be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 770.681108] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e708e6-3fc6-9501-f74e-05f3448b8e32/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 770.681326] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e708e6-3fc6-9501-f74e-05f3448b8e32/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 770.753685] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d67984-c2ca-51e3-2194-674c04468257, 'name': SearchDatastore_Task, 'duration_secs': 0.009864} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.754012] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.754405] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.754666] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.754828] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.755059] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.755379] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48a0f9cc-4567-419a-b080-018778818690 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.763879] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.766762] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.766941] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.767919] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-423f8d4b-dc18-48c5-a0cd-dc94ed2abd0f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.773688] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 770.773688] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f79007-757c-f916-d661-c6e049539080" [ 770.773688] env[61991]: _type = "Task" [ 770.773688] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.781719] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f79007-757c-f916-d661-c6e049539080, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.783591] env[61991]: INFO nova.compute.manager [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] instance snapshotting [ 770.784223] env[61991]: DEBUG nova.objects.instance [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'flavor' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 770.796522] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2c425356-d3ec-4940-b890-574c9d0ec7ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.886059] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129635, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.924953] env[61991]: DEBUG nova.network.neutron [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Updated VIF entry in instance network info cache for port 6a28231d-a28f-49ea-8339-3ca4a02684bb. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.925367] env[61991]: DEBUG nova.network.neutron [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Updating instance_info_cache with network_info: [{"id": "6a28231d-a28f-49ea-8339-3ca4a02684bb", "address": "fa:16:3e:15:a3:26", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a28231d-a2", "ovs_interfaceid": "6a28231d-a28f-49ea-8339-3ca4a02684bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.143361] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.143661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.143889] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.144643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.144883] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.147394] env[61991]: INFO nova.compute.manager [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Terminating instance [ 771.158895] env[61991]: DEBUG nova.compute.manager [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.158895] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.158895] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a63f76a-25a1-4000-a9c2-ce583a0f05c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.162794] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 771.163498] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d9f6ac3-3c97-4510-a979-290fbff61555 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.172625] env[61991]: DEBUG oslo_vmware.api [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 771.172625] env[61991]: value = "task-1129637" [ 771.172625] env[61991]: _type = "Task" [ 771.172625] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.186179] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance e3e5ec72-eec1-4155-bab2-038e3d5e38a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.187920] env[61991]: DEBUG oslo_vmware.api [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129637, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.289678] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f79007-757c-f916-d661-c6e049539080, 'name': SearchDatastore_Task, 'duration_secs': 0.009993} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.290730] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.292200] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5f546cf-dbb9-4008-b56e-fb7eb0da6c55 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.295131] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf900e72-045b-489c-b457-f56e1cbf719d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.301379] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 771.301379] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522ce871-e79f-15ba-6965-901e03a6a450" [ 771.301379] env[61991]: _type = "Task" [ 771.301379] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.323398] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eeb02a3-8635-40ca-bbae-fc835a38dfc4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.338942] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522ce871-e79f-15ba-6965-901e03a6a450, 'name': SearchDatastore_Task, 'duration_secs': 0.012231} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.340077] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.340240] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 0aa67bb1-59d3-42e4-8f12-47e61a130dfc/0aa67bb1-59d3-42e4-8f12-47e61a130dfc.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.340501] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4277a72c-9a7c-4db7-9f8c-5d5f67c0880f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.348584] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 771.348584] env[61991]: value = "task-1129638" [ 771.348584] env[61991]: _type = "Task" [ 771.348584] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.359046] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129638, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.390570] env[61991]: DEBUG oslo_vmware.api [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129635, 'name': PowerOnVM_Task, 'duration_secs': 0.998999} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.390899] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 771.391247] env[61991]: INFO nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Took 9.07 seconds to spawn the instance on the hypervisor. [ 771.391521] env[61991]: DEBUG nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 771.392882] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22993f2-544f-4739-981f-bac31f1a3f27 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.429930] env[61991]: DEBUG oslo_concurrency.lockutils [req-e45c1998-453b-463e-99e6-f7e4db7c7987 req-0a75503d-5d60-438c-a933-4a19b36aa612 service nova] Releasing lock "refresh_cache-0aa67bb1-59d3-42e4-8f12-47e61a130dfc" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.659558] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.690947] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3e62f219-7fec-470f-9d0a-e292d62e076c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.693077] env[61991]: DEBUG oslo_vmware.api [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129637, 'name': PowerOffVM_Task, 'duration_secs': 0.319684} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.697059] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.697421] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.697757] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.698147] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.698356] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.698724] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.699544] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.699998] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.700631] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.700631] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.702093] env[61991]: DEBUG nova.virt.hardware [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.703504] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 771.704397] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 771.706364] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35a7029-f2a0-42a0-8118-58bfd1764283 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.709919] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f36bc0bb-7e0c-46ff-a87f-0aea5c84f6b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.721483] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6290546-6b1a-4ded-9d71-dd63129b6286 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.777786] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 771.778509] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 771.778642] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleting the datastore file [datastore1] d2a0e5d1-5557-41b6-936d-dc86d1346c61 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.779478] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4175c027-5fc7-42f4-800f-d77228e14311 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.785933] env[61991]: DEBUG oslo_vmware.api [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 771.785933] env[61991]: value = "task-1129640" [ 771.785933] env[61991]: _type = "Task" [ 771.785933] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.797023] env[61991]: DEBUG oslo_vmware.api [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.841541] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 771.842011] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f7acac11-6515-43e4-b544-70ac9b929188 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.851125] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 771.851125] env[61991]: value = "task-1129641" [ 771.851125] env[61991]: _type = "Task" [ 771.851125] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.867963] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129641, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.871659] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129638, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.917770] env[61991]: INFO nova.compute.manager [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Took 39.44 seconds to build instance. [ 772.210027] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 884e465b-e14f-4114-81a6-40e2e8694558 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.297539] env[61991]: DEBUG oslo_vmware.api [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.261962} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.297970] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.298258] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 772.298547] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.298817] env[61991]: INFO nova.compute.manager [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Took 1.15 seconds to destroy the instance on the hypervisor. [ 772.299110] env[61991]: DEBUG oslo.service.loopingcall [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.299801] env[61991]: DEBUG nova.compute.manager [-] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.299909] env[61991]: DEBUG nova.network.neutron [-] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.364164] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129638, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567625} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.364667] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 0aa67bb1-59d3-42e4-8f12-47e61a130dfc/0aa67bb1-59d3-42e4-8f12-47e61a130dfc.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 772.365337] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.366611] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a893ca49-97ca-43ee-9bd3-c0d8555594e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.375931] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129641, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.384433] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 772.384433] env[61991]: value = "task-1129642" [ 772.384433] env[61991]: _type = "Task" [ 772.384433] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.395282] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.422111] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dfb3cd16-e420-4fd6-8c31-23829dffc86d tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.656s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.527820] env[61991]: DEBUG nova.compute.manager [req-c2601e14-ac7e-46c7-afc3-e0f71b34d9ec req-48842a3e-de6c-4506-8736-4da8e9a23fb0 service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Received event network-vif-plugged-2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.528065] env[61991]: DEBUG oslo_concurrency.lockutils [req-c2601e14-ac7e-46c7-afc3-e0f71b34d9ec req-48842a3e-de6c-4506-8736-4da8e9a23fb0 service nova] Acquiring lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.528279] env[61991]: DEBUG oslo_concurrency.lockutils [req-c2601e14-ac7e-46c7-afc3-e0f71b34d9ec req-48842a3e-de6c-4506-8736-4da8e9a23fb0 service nova] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.528453] env[61991]: DEBUG oslo_concurrency.lockutils [req-c2601e14-ac7e-46c7-afc3-e0f71b34d9ec req-48842a3e-de6c-4506-8736-4da8e9a23fb0 service nova] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.528710] env[61991]: DEBUG nova.compute.manager [req-c2601e14-ac7e-46c7-afc3-e0f71b34d9ec req-48842a3e-de6c-4506-8736-4da8e9a23fb0 service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] No waiting events found dispatching network-vif-plugged-2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 772.528926] env[61991]: WARNING nova.compute.manager [req-c2601e14-ac7e-46c7-afc3-e0f71b34d9ec req-48842a3e-de6c-4506-8736-4da8e9a23fb0 service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Received unexpected event network-vif-plugged-2d6d64e8-636f-4478-93ab-f1ca63a74ce1 for instance with vm_state building and task_state spawning. [ 772.677338] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Successfully updated port: 2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 772.713365] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f6945e87-3b17-4e0e-845f-3bacf0fbe5ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.867840] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129641, 'name': CreateSnapshot_Task, 'duration_secs': 0.701942} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.867840] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 772.869566] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c410c2-f7fc-4693-bdf4-de6bff679c87 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.898166] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090921} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.898715] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.899651] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ab5fe4-f79c-448d-94de-4c230fa0e06b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.923255] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 0aa67bb1-59d3-42e4-8f12-47e61a130dfc/0aa67bb1-59d3-42e4-8f12-47e61a130dfc.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.925014] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae3fcd3c-1acf-41c1-9dc6-a6d938df8544 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.949806] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 772.949806] env[61991]: value = "task-1129643" [ 772.949806] env[61991]: _type = "Task" [ 772.949806] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.958284] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129643, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.179965] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "refresh_cache-c279a82f-bf4c-4a59-b04a-38d31070b0e5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.180153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "refresh_cache-c279a82f-bf4c-4a59-b04a-38d31070b0e5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.180309] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 773.216311] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 49324fef-ad48-451b-a5ce-d9a1231137db has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.256033] env[61991]: DEBUG nova.network.neutron [-] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.391170] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 773.392304] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-585a7628-d018-4a42-85de-f01fa36ef1c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.409042] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 773.409042] env[61991]: value = "task-1129644" [ 773.409042] env[61991]: _type = "Task" [ 773.409042] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.417748] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129644, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.457065] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129643, 'name': ReconfigVM_Task, 'duration_secs': 0.472725} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.457065] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 0aa67bb1-59d3-42e4-8f12-47e61a130dfc/0aa67bb1-59d3-42e4-8f12-47e61a130dfc.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.457816] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abe210bc-faf9-4dc3-860b-b1541aaeb436 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.465135] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 773.465135] env[61991]: value = "task-1129645" [ 773.465135] env[61991]: _type = "Task" [ 773.465135] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.472996] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129645, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.719445] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 62b3e5ae-b90c-47f8-95c4-14587dbf647d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.743125] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.762028] env[61991]: INFO nova.compute.manager [-] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Took 1.46 seconds to deallocate network for instance. [ 773.917481] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129644, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.949055] env[61991]: DEBUG nova.network.neutron [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Updating instance_info_cache with network_info: [{"id": "2d6d64e8-636f-4478-93ab-f1ca63a74ce1", "address": "fa:16:3e:7f:e2:73", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d6d64e8-63", "ovs_interfaceid": "2d6d64e8-636f-4478-93ab-f1ca63a74ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.978597] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129645, 'name': Rename_Task, 'duration_secs': 0.306517} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.979072] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.979346] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9917f1de-c726-4642-9f6b-9c3de533824d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.987104] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 773.987104] env[61991]: value = "task-1129646" [ 773.987104] env[61991]: _type = "Task" [ 773.987104] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.997287] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.227031] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d36bd981-9ea3-46f6-8376-ac1e0c3bf61e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.227031] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 774.227031] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3456MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 774.274072] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.416662] env[61991]: DEBUG nova.compute.manager [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Received event network-changed-179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.417442] env[61991]: DEBUG nova.compute.manager [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Refreshing instance network info cache due to event network-changed-179a25d5-af58-45b5-b510-6f9064bddb3c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 774.417442] env[61991]: DEBUG oslo_concurrency.lockutils [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] Acquiring lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.417561] env[61991]: DEBUG oslo_concurrency.lockutils [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] Acquired lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.418125] env[61991]: DEBUG nova.network.neutron [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Refreshing network info cache for port 179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 774.426615] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129644, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.449239] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "refresh_cache-c279a82f-bf4c-4a59-b04a-38d31070b0e5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.449562] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Instance network_info: |[{"id": "2d6d64e8-636f-4478-93ab-f1ca63a74ce1", "address": "fa:16:3e:7f:e2:73", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d6d64e8-63", "ovs_interfaceid": "2d6d64e8-636f-4478-93ab-f1ca63a74ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 774.450280] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:e2:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d6d64e8-636f-4478-93ab-f1ca63a74ce1', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 774.458461] env[61991]: DEBUG oslo.service.loopingcall [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.461581] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 774.462193] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54a75e2d-ea34-44f6-868e-0ac7f39db678 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.490361] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 774.490361] env[61991]: value = "task-1129647" [ 774.490361] env[61991]: _type = "Task" [ 774.490361] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.504449] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129646, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.508615] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129647, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.533195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.533195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.533195] env[61991]: DEBUG nova.objects.instance [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'flavor' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 774.658188] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd5d9d9-5f31-416a-aae4-c93d82bac7f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.667311] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7d7f6e-d0d0-48f8-8c79-20fbaf2cb7a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.700451] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33671c97-042f-49a6-b1e0-2babc3a90791 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.708533] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22493a18-e0c1-49e2-9911-ccb2ea9730b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.722532] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 774.924021] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129644, 'name': CloneVM_Task, 'duration_secs': 1.316528} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.924021] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created linked-clone VM from snapshot [ 774.924460] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65745dbc-c536-4555-9294-9bd52962f0a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.934545] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Uploading image 5b317551-bad7-4685-b760-e9c291dff08c {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 774.956120] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 774.956120] env[61991]: value = "vm-246859" [ 774.956120] env[61991]: _type = "VirtualMachine" [ 774.956120] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 774.957609] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e13e389e-b448-4252-b539-530c41506ae5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.964834] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease: (returnval){ [ 774.964834] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525de73b-3294-ac94-afef-6086f4efe9a3" [ 774.964834] env[61991]: _type = "HttpNfcLease" [ 774.964834] env[61991]: } obtained for exporting VM: (result){ [ 774.964834] env[61991]: value = "vm-246859" [ 774.964834] env[61991]: _type = "VirtualMachine" [ 774.964834] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 774.965483] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the lease: (returnval){ [ 774.965483] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525de73b-3294-ac94-afef-6086f4efe9a3" [ 774.965483] env[61991]: _type = "HttpNfcLease" [ 774.965483] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 774.972406] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 774.972406] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525de73b-3294-ac94-afef-6086f4efe9a3" [ 774.972406] env[61991]: _type = "HttpNfcLease" [ 774.972406] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 775.003521] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129646, 'name': PowerOnVM_Task, 'duration_secs': 0.663586} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.003931] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 775.004259] env[61991]: INFO nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Took 8.06 seconds to spawn the instance on the hypervisor. [ 775.004570] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 775.008176] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df31f323-7ff3-4df9-9aa7-689f3422631d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.010563] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129647, 'name': CreateVM_Task, 'duration_secs': 0.404441} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.011719] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 775.012791] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.012980] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.013416] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 775.013741] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d37a18b-7fb8-4343-bdde-0d58401ba0e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.023148] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 775.023148] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5215d07f-7c88-f86e-625e-ceea5d0e46c0" [ 775.023148] env[61991]: _type = "Task" [ 775.023148] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.031589] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5215d07f-7c88-f86e-625e-ceea5d0e46c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.046072] env[61991]: DEBUG nova.objects.instance [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'pci_requests' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 775.161493] env[61991]: DEBUG nova.network.neutron [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updated VIF entry in instance network info cache for port 179a25d5-af58-45b5-b510-6f9064bddb3c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 775.161887] env[61991]: DEBUG nova.network.neutron [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updating instance_info_cache with network_info: [{"id": "179a25d5-af58-45b5-b510-6f9064bddb3c", "address": "fa:16:3e:30:74:a2", "network": {"id": "4dff2d72-8dde-4106-bd50-da1177efb4bd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1324618604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9eb579be6d0d439db8465cc9639bdffb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap179a25d5-af", "ovs_interfaceid": "179a25d5-af58-45b5-b510-6f9064bddb3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.258324] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 775.258597] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 81 to 82 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 775.258887] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 775.441267] env[61991]: DEBUG nova.compute.manager [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Received event network-changed-2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.441471] env[61991]: DEBUG nova.compute.manager [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Refreshing instance network info cache due to event network-changed-2d6d64e8-636f-4478-93ab-f1ca63a74ce1. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 775.441687] env[61991]: DEBUG oslo_concurrency.lockutils [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] Acquiring lock "refresh_cache-c279a82f-bf4c-4a59-b04a-38d31070b0e5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.441874] env[61991]: DEBUG oslo_concurrency.lockutils [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] Acquired lock "refresh_cache-c279a82f-bf4c-4a59-b04a-38d31070b0e5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.441985] env[61991]: DEBUG nova.network.neutron [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Refreshing network info cache for port 2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 775.473435] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 775.473435] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525de73b-3294-ac94-afef-6086f4efe9a3" [ 775.473435] env[61991]: _type = "HttpNfcLease" [ 775.473435] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 775.473737] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 775.473737] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525de73b-3294-ac94-afef-6086f4efe9a3" [ 775.473737] env[61991]: _type = "HttpNfcLease" [ 775.473737] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 775.474492] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e2199b-e784-4b1c-81af-c03db83fc6fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.482214] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9855-8428-2f12-1193-c18147daeb18/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 775.482446] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9855-8428-2f12-1193-c18147daeb18/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 775.551024] env[61991]: DEBUG nova.objects.base [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Object Instance<80d05278-b516-4408-94b0-11bc93500b5c> lazy-loaded attributes: flavor,pci_requests {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 775.551264] env[61991]: DEBUG nova.network.neutron [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.553269] env[61991]: INFO nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Took 38.87 seconds to build instance. [ 775.561265] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5215d07f-7c88-f86e-625e-ceea5d0e46c0, 'name': SearchDatastore_Task, 'duration_secs': 0.016904} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.561578] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.561888] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.562222] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.562412] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.562601] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.562866] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37920c98-8a40-4c34-bfe5-c255a68516db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.571574] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.571757] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 775.572749] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b65ef022-41e8-404c-bbf7-0905be33e0eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.578272] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 775.578272] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fb9ff9-3e4c-e550-7c4b-b73f1d7ca5f6" [ 775.578272] env[61991]: _type = "Task" [ 775.578272] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.586660] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fb9ff9-3e4c-e550-7c4b-b73f1d7ca5f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.621672] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fcfaef0b-d241-42ab-866e-8090aaef9e41 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.637458] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0a43a66b-124a-40c5-80cd-a6f7115bbc94 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.105s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.664930] env[61991]: DEBUG oslo_concurrency.lockutils [req-95abca54-a47e-48bd-8d62-cb9ebf4dcab2 req-721de304-be9c-4130-9760-28f646b16573 service nova] Releasing lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.764308] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 775.764710] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.151s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.765132] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.208s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.765338] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.767163] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.209s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.770501] env[61991]: INFO nova.compute.claims [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.773683] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 775.773683] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Cleaning up deleted instances {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 775.799897] env[61991]: INFO nova.scheduler.client.report [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Deleted allocations for instance acef131a-4263-42ad-964e-bb1e0cb21eec [ 776.055952] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.821s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.088964] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fb9ff9-3e4c-e550-7c4b-b73f1d7ca5f6, 'name': SearchDatastore_Task, 'duration_secs': 0.013537} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.090180] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4ccc69d-8f1c-4ceb-af61-8df3f757fe9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.098799] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 776.098799] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba79e4-304f-72e5-2343-94bc3d9a3605" [ 776.098799] env[61991]: _type = "Task" [ 776.098799] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.107595] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba79e4-304f-72e5-2343-94bc3d9a3605, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.173744] env[61991]: DEBUG nova.network.neutron [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Updated VIF entry in instance network info cache for port 2d6d64e8-636f-4478-93ab-f1ca63a74ce1. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 776.174282] env[61991]: DEBUG nova.network.neutron [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Updating instance_info_cache with network_info: [{"id": "2d6d64e8-636f-4478-93ab-f1ca63a74ce1", "address": "fa:16:3e:7f:e2:73", "network": {"id": "7772fa64-9110-4a7b-8121-8964846ac8ec", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1675729695-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fd3b8e07dce44a8b03543c3c0ca1bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d6d64e8-63", "ovs_interfaceid": "2d6d64e8-636f-4478-93ab-f1ca63a74ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.282147] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] There are 14 instances to clean {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 776.282909] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 30bd843a-9d86-4de1-b011-600e2a93bcb3] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 776.310044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-843441c9-7bb2-4981-9d39-86f20b182ce2 tempest-ServerShowV247Test-844570035 tempest-ServerShowV247Test-844570035-project-member] Lock "acef131a-4263-42ad-964e-bb1e0cb21eec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.534s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.610095] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba79e4-304f-72e5-2343-94bc3d9a3605, 'name': SearchDatastore_Task, 'duration_secs': 0.01276} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.610638] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.611097] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] c279a82f-bf4c-4a59-b04a-38d31070b0e5/c279a82f-bf4c-4a59-b04a-38d31070b0e5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 776.611630] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17c90110-2192-4f10-af3e-31e70fda2d37 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.619056] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 776.619056] env[61991]: value = "task-1129649" [ 776.619056] env[61991]: _type = "Task" [ 776.619056] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.627565] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129649, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.678115] env[61991]: DEBUG oslo_concurrency.lockutils [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] Releasing lock "refresh_cache-c279a82f-bf4c-4a59-b04a-38d31070b0e5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.678598] env[61991]: DEBUG nova.compute.manager [req-3f5131cf-ec19-4457-9842-400f89578b5e req-53d55370-120e-46f1-8f32-b8e874808dba service nova] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Received event network-vif-deleted-46746327-a6c2-4236-b0eb-823d21d735a4 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.787215] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 0631634f-5c58-4ab0-b0dd-71821caf2fc6] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 777.134884] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129649, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.291109] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 43cb61e7-1748-40d9-a287-1179c8219c2a] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 777.346555] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6453e220-a88a-491d-91e8-225be277eec0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.355788] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc94889-1dc8-4286-8c18-1684b9e09a01 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.390054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907972c2-7f3b-46eb-8507-932ad23003c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.399862] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b6d650-cd18-4474-a2c5-849b05897720 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.413770] env[61991]: DEBUG nova.compute.provider_tree [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.631575] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129649, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613914} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.632062] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] c279a82f-bf4c-4a59-b04a-38d31070b0e5/c279a82f-bf4c-4a59-b04a-38d31070b0e5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 777.632308] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 777.632685] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1cf74555-ec59-4bb3-9132-5a5534e2819e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.639439] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 777.639439] env[61991]: value = "task-1129650" [ 777.639439] env[61991]: _type = "Task" [ 777.639439] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.648334] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129650, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.794454] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 527561af-ba19-447f-a4ee-618824b9ce0f] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 777.917596] env[61991]: DEBUG nova.scheduler.client.report [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.958053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.958378] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.958758] env[61991]: DEBUG nova.objects.instance [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'flavor' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.149810] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129650, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097118} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.150155] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.150944] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d4b19d-1895-4fe5-8b78-43c7b30cc7c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.174982] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] c279a82f-bf4c-4a59-b04a-38d31070b0e5/c279a82f-bf4c-4a59-b04a-38d31070b0e5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.175477] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21b7edce-11e6-440d-a896-37e2a6729c0c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.198764] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 778.198764] env[61991]: value = "task-1129651" [ 778.198764] env[61991]: _type = "Task" [ 778.198764] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.209139] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129651, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.297679] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: fe78a8de-6174-47b3-9271-171c4ba76068] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 778.424061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.424483] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.427808] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 28.018s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.470144] env[61991]: DEBUG nova.compute.manager [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Received event network-changed-179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.470468] env[61991]: DEBUG nova.compute.manager [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Refreshing instance network info cache due to event network-changed-179a25d5-af58-45b5-b510-6f9064bddb3c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.470636] env[61991]: DEBUG oslo_concurrency.lockutils [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] Acquiring lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.470821] env[61991]: DEBUG oslo_concurrency.lockutils [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] Acquired lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.471031] env[61991]: DEBUG nova.network.neutron [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Refreshing network info cache for port 179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.549897] env[61991]: DEBUG nova.objects.instance [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'pci_requests' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.709811] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129651, 'name': ReconfigVM_Task, 'duration_secs': 0.440429} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.710162] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Reconfigured VM instance instance-0000003f to attach disk [datastore2] c279a82f-bf4c-4a59-b04a-38d31070b0e5/c279a82f-bf4c-4a59-b04a-38d31070b0e5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.710882] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb524144-1927-4b15-b202-5d74533092d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.717997] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 778.717997] env[61991]: value = "task-1129652" [ 778.717997] env[61991]: _type = "Task" [ 778.717997] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.730480] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129652, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.801408] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 812d3491-0598-4171-b905-5304ebf6ae37] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 778.930744] env[61991]: DEBUG nova.objects.instance [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lazy-loading 'migration_context' on Instance uuid c14d99a3-950d-44d0-b330-3031a1f8a2be {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.933682] env[61991]: DEBUG nova.compute.utils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.935983] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.936214] env[61991]: DEBUG nova.network.neutron [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.982367] env[61991]: DEBUG nova.policy [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a385cf261940a19eb17046958920ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3015571d285418aae0fca50b563394e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.052618] env[61991]: DEBUG nova.objects.base [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Object Instance<80d05278-b516-4408-94b0-11bc93500b5c> lazy-loaded attributes: flavor,pci_requests {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 779.052940] env[61991]: DEBUG nova.network.neutron [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 779.093444] env[61991]: DEBUG nova.policy [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.231899] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129652, 'name': Rename_Task, 'duration_secs': 0.188962} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.235041] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 779.235041] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-79ac6b9d-4235-465f-8d5b-27878ecfc057 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.240455] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 779.240455] env[61991]: value = "task-1129653" [ 779.240455] env[61991]: _type = "Task" [ 779.240455] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.251994] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129653, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.308232] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9d88002b-d5f0-4669-9a34-bfbf8f9169d3] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 779.339130] env[61991]: DEBUG nova.network.neutron [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updated VIF entry in instance network info cache for port 179a25d5-af58-45b5-b510-6f9064bddb3c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.339130] env[61991]: DEBUG nova.network.neutron [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updating instance_info_cache with network_info: [{"id": "179a25d5-af58-45b5-b510-6f9064bddb3c", "address": "fa:16:3e:30:74:a2", "network": {"id": "4dff2d72-8dde-4106-bd50-da1177efb4bd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1324618604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9eb579be6d0d439db8465cc9639bdffb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap179a25d5-af", "ovs_interfaceid": "179a25d5-af58-45b5-b510-6f9064bddb3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.420896] env[61991]: DEBUG nova.network.neutron [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Successfully created port: a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.435878] env[61991]: DEBUG nova.network.neutron [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Successfully created port: 2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.441147] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.752548] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129653, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.809900] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: a2606541-ff39-4d03-bd19-d0ee8e035c08] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 779.840467] env[61991]: DEBUG oslo_concurrency.lockutils [req-daeb8db1-57d0-4375-8ef8-45c6e7bec5d9 req-d3b954b9-d6d8-40b0-9f20-ca83be93e382 service nova] Releasing lock "refresh_cache-d6a2d226-5097-4461-b2ce-fa698f1066d5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.898996] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae602672-45fd-45f4-bb2d-560e441cc7c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.907393] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2811bb-e696-475c-92b8-457b9cfb4a12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.943247] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1982e01-f5a8-41e0-8bcb-de4a67b9a23c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.954889] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0e2ab3-3c07-4631-a438-bc469b9a985a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.971598] env[61991]: DEBUG nova.compute.provider_tree [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.254016] env[61991]: DEBUG oslo_vmware.api [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129653, 'name': PowerOnVM_Task, 'duration_secs': 0.683819} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.254352] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 780.254580] env[61991]: INFO nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Took 8.59 seconds to spawn the instance on the hypervisor. [ 780.254778] env[61991]: DEBUG nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.255696] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ace0afd-d1b6-4a57-a211-094fc6875f66 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.314032] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: b1c1c4e4-a82b-4066-8ee8-342d5ad18d29] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 780.451230] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.474147] env[61991]: DEBUG nova.scheduler.client.report [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.774058] env[61991]: INFO nova.compute.manager [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Took 41.14 seconds to build instance. [ 780.816674] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 598c0bd7-baf5-4d77-8d06-0f83a08b9685] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 780.986394] env[61991]: DEBUG nova.network.neutron [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Successfully updated port: 2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.259474] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='8db13f037266d1afdfe111b122d3138e',container_format='bare',created_at=2024-10-04T09:20:44Z,direct_url=,disk_format='vmdk',id=e4b0de21-26e3-4d65-834b-7a893af37f7a,min_disk=1,min_ram=0,name='tempest-test-snap-1271349021',owner='e3015571d285418aae0fca50b563394e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-04T09:20:59Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.259792] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.259976] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.260100] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.260252] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.260462] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.260621] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.260783] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.260952] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.261133] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.261310] env[61991]: DEBUG nova.virt.hardware [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.264510] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593075a1-c08f-4758-a650-a2cdd002a431 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.272839] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cbd16b-d543-44a4-acde-206cc5dd75ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.278964] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e708e6-3fc6-9501-f74e-05f3448b8e32/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 781.279523] env[61991]: DEBUG oslo_concurrency.lockutils [None req-29c0ab2b-b656-4ab6-b0ff-2309deb10c00 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.013s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.280234] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5fdb0f-6f58-4ab7-b20b-b37f851cd573 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.286349] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e708e6-3fc6-9501-f74e-05f3448b8e32/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 781.286530] env[61991]: ERROR oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e708e6-3fc6-9501-f74e-05f3448b8e32/disk-0.vmdk due to incomplete transfer. [ 781.294306] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ce00a3f6-9bbb-4990-8f3b-90203fbfc930 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.302205] env[61991]: DEBUG oslo_vmware.rw_handles [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e708e6-3fc6-9501-f74e-05f3448b8e32/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 781.302422] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Uploaded image 0810354d-ca0b-4776-8487-42797b713043 to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 781.304673] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 781.304929] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dd8f83df-fd47-4b5d-b3a3-d5d60f043b07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.310965] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 781.310965] env[61991]: value = "task-1129654" [ 781.310965] env[61991]: _type = "Task" [ 781.310965] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.318701] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129654, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.320204] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 2866aec9-56fd-4e77-b651-9d85783fb94f] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 781.337416] env[61991]: DEBUG nova.network.neutron [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Successfully updated port: a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.473464] env[61991]: DEBUG nova.compute.manager [req-271f7a18-58f0-4669-adb9-0e93abc00344 req-faaeca92-c9b6-4c3a-b867-c846f1e7f501 service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Received event network-vif-plugged-2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.473567] env[61991]: DEBUG oslo_concurrency.lockutils [req-271f7a18-58f0-4669-adb9-0e93abc00344 req-faaeca92-c9b6-4c3a-b867-c846f1e7f501 service nova] Acquiring lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.473763] env[61991]: DEBUG oslo_concurrency.lockutils [req-271f7a18-58f0-4669-adb9-0e93abc00344 req-faaeca92-c9b6-4c3a-b867-c846f1e7f501 service nova] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.473882] env[61991]: DEBUG oslo_concurrency.lockutils [req-271f7a18-58f0-4669-adb9-0e93abc00344 req-faaeca92-c9b6-4c3a-b867-c846f1e7f501 service nova] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.474077] env[61991]: DEBUG nova.compute.manager [req-271f7a18-58f0-4669-adb9-0e93abc00344 req-faaeca92-c9b6-4c3a-b867-c846f1e7f501 service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] No waiting events found dispatching network-vif-plugged-2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.474263] env[61991]: WARNING nova.compute.manager [req-271f7a18-58f0-4669-adb9-0e93abc00344 req-faaeca92-c9b6-4c3a-b867-c846f1e7f501 service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Received unexpected event network-vif-plugged-2d37704c-b54b-459f-b593-12a9062c3601 for instance with vm_state building and task_state spawning. [ 781.484336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.056s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.490966] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.213s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.492536] env[61991]: INFO nova.compute.claims [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.495464] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-e3e5ec72-eec1-4155-bab2-038e3d5e38a8" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.495610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-e3e5ec72-eec1-4155-bab2-038e3d5e38a8" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.495854] env[61991]: DEBUG nova.network.neutron [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.597839] env[61991]: DEBUG nova.compute.manager [req-57e66320-514d-481a-80ad-7c013e8e4f2c req-41400241-5346-41e5-8eab-ebefddb9237d service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-vif-plugged-a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.598080] env[61991]: DEBUG oslo_concurrency.lockutils [req-57e66320-514d-481a-80ad-7c013e8e4f2c req-41400241-5346-41e5-8eab-ebefddb9237d service nova] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.598295] env[61991]: DEBUG oslo_concurrency.lockutils [req-57e66320-514d-481a-80ad-7c013e8e4f2c req-41400241-5346-41e5-8eab-ebefddb9237d service nova] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.598467] env[61991]: DEBUG oslo_concurrency.lockutils [req-57e66320-514d-481a-80ad-7c013e8e4f2c req-41400241-5346-41e5-8eab-ebefddb9237d service nova] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.598674] env[61991]: DEBUG nova.compute.manager [req-57e66320-514d-481a-80ad-7c013e8e4f2c req-41400241-5346-41e5-8eab-ebefddb9237d service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] No waiting events found dispatching network-vif-plugged-a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.598912] env[61991]: WARNING nova.compute.manager [req-57e66320-514d-481a-80ad-7c013e8e4f2c req-41400241-5346-41e5-8eab-ebefddb9237d service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received unexpected event network-vif-plugged-a20e94c9-8ee3-4696-b020-423d80f00dc9 for instance with vm_state active and task_state None. [ 781.821372] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129654, 'name': Destroy_Task, 'duration_secs': 0.325117} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.821650] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Destroyed the VM [ 781.821887] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 781.822168] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9930ca99-b73b-4413-a165-fee1d72f0d96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.827799] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 432ab035-b5b0-4390-bef2-ec2691b4422b] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 781.831614] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 781.831614] env[61991]: value = "task-1129655" [ 781.831614] env[61991]: _type = "Task" [ 781.831614] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.840162] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.840357] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.840542] env[61991]: DEBUG nova.network.neutron [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.842616] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129655, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.854467] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "d6a2d226-5097-4461-b2ce-fa698f1066d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.854830] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.855084] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "d6a2d226-5097-4461-b2ce-fa698f1066d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.855282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.855454] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.858126] env[61991]: INFO nova.compute.manager [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Terminating instance [ 781.861174] env[61991]: DEBUG nova.compute.manager [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 781.861174] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.861726] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc69d041-0483-490c-8c13-5d54bef602ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.870110] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.870393] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a442401-b49a-4df0-911e-c54064590ffb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.876432] env[61991]: DEBUG oslo_vmware.api [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 781.876432] env[61991]: value = "task-1129656" [ 781.876432] env[61991]: _type = "Task" [ 781.876432] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.885335] env[61991]: DEBUG oslo_vmware.api [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129656, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.057072] env[61991]: DEBUG nova.network.neutron [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.265287] env[61991]: DEBUG nova.network.neutron [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Updating instance_info_cache with network_info: [{"id": "2d37704c-b54b-459f-b593-12a9062c3601", "address": "fa:16:3e:a3:e5:75", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d37704c-b5", "ovs_interfaceid": "2d37704c-b54b-459f-b593-12a9062c3601", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.330597] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c2ef2717-4eac-428b-92a8-12a37d7c0719] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 782.344893] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129655, 'name': RemoveSnapshot_Task, 'duration_secs': 0.45507} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.345564] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 782.346370] env[61991]: DEBUG nova.compute.manager [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.347037] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bed53eb-988e-4a16-95a3-66334f18ca16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.375268] env[61991]: WARNING nova.network.neutron [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] e6f9d471-4f42-4781-93c3-f4a9ff84317b already exists in list: networks containing: ['e6f9d471-4f42-4781-93c3-f4a9ff84317b']. ignoring it [ 782.390501] env[61991]: DEBUG oslo_vmware.api [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129656, 'name': PowerOffVM_Task, 'duration_secs': 0.178642} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.390838] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 782.391094] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 782.391785] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8afd6671-2e7a-4cc4-b9a3-f47cce1797fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.452258] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 782.452492] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 782.452674] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Deleting the datastore file [datastore2] d6a2d226-5097-4461-b2ce-fa698f1066d5 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.452932] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01a7b2ce-44fa-4bb7-abe8-eb2a53183d8a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.459109] env[61991]: DEBUG oslo_vmware.api [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for the task: (returnval){ [ 782.459109] env[61991]: value = "task-1129658" [ 782.459109] env[61991]: _type = "Task" [ 782.459109] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.467738] env[61991]: DEBUG oslo_vmware.api [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.771239] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-e3e5ec72-eec1-4155-bab2-038e3d5e38a8" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.771579] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Instance network_info: |[{"id": "2d37704c-b54b-459f-b593-12a9062c3601", "address": "fa:16:3e:a3:e5:75", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d37704c-b5", "ovs_interfaceid": "2d37704c-b54b-459f-b593-12a9062c3601", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.772499] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:e5:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3952eb02-1162-48ed-8227-9c138960d583', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d37704c-b54b-459f-b593-12a9062c3601', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.779551] env[61991]: DEBUG oslo.service.loopingcall [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.783044] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.783950] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6075509-6150-4d0c-939a-a5fc7fa73b1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.805980] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.805980] env[61991]: value = "task-1129659" [ 782.805980] env[61991]: _type = "Task" [ 782.805980] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.811682] env[61991]: DEBUG nova.network.neutron [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "address": "fa:16:3e:5c:e6:a5", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20e94c9-8e", "ovs_interfaceid": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.819726] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129659, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.836813] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: d978d68a-0709-4a32-bbaf-3bfa006be85c] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 782.860236] env[61991]: INFO nova.compute.manager [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Shelve offloading [ 782.865150] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.865466] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02e78cec-133c-4e45-a595-d3a0480239d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.876161] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 782.876161] env[61991]: value = "task-1129660" [ 782.876161] env[61991]: _type = "Task" [ 782.876161] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.889168] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 782.889168] env[61991]: DEBUG nova.compute.manager [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.890328] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4672854b-e965-4dd8-98e4-a790f25db97c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.897227] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.897433] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.897669] env[61991]: DEBUG nova.network.neutron [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.959242] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b4ba42-eeb8-4703-926c-7e6e20162d89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.973032] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208349fb-99d8-4090-ab13-cc19fd6b8a06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.978371] env[61991]: DEBUG oslo_vmware.api [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Task: {'id': task-1129658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199878} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.978793] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.979041] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.979460] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.979522] env[61991]: INFO nova.compute.manager [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 782.980372] env[61991]: DEBUG oslo.service.loopingcall [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.980471] env[61991]: DEBUG nova.compute.manager [-] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.980518] env[61991]: DEBUG nova.network.neutron [-] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.008013] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea0457c-4174-4b5d-b74a-a557c246f4bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.017021] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2028611b-70ba-4127-9694-7d21517acdf0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.030714] env[61991]: DEBUG nova.compute.provider_tree [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.056620] env[61991]: INFO nova.compute.manager [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Swapping old allocation on dict_keys(['d748992a-e0bf-4ec2-9c17-0e373360e5a3']) held by migration 528ff66f-4caf-4189-ac3a-ace614777b67 for instance [ 783.084648] env[61991]: DEBUG nova.scheduler.client.report [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Overwriting current allocation {'allocations': {'d748992a-e0bf-4ec2-9c17-0e373360e5a3': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 82}}, 'project_id': '518b363cb2db43c288d25c39a239c9c1', 'user_id': 'f2666f4b636d4e4d8682a5876772d381', 'consumer_generation': 1} on consumer c14d99a3-950d-44d0-b330-3031a1f8a2be {{(pid=61991) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 783.176623] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.176826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquired lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.177015] env[61991]: DEBUG nova.network.neutron [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 783.239035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.239035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.239300] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.239386] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.239591] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.241873] env[61991]: INFO nova.compute.manager [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Terminating instance [ 783.244015] env[61991]: DEBUG nova.compute.manager [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.244221] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.245121] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c815305-95d8-4df2-9e58-ba158f9c2e7e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.253139] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.253684] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-faac1951-0918-4aef-939f-38a159f77402 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.260928] env[61991]: DEBUG oslo_vmware.api [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 783.260928] env[61991]: value = "task-1129661" [ 783.260928] env[61991]: _type = "Task" [ 783.260928] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.272691] env[61991]: DEBUG oslo_vmware.api [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129661, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.317751] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129659, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.324201] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.324842] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.325195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.326727] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b60552-f26b-4ad1-8fb1-1fbce8cb85f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.350597] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 783.351012] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 783.351267] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.351524] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 783.351740] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.352386] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 783.352386] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 783.352386] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 783.352617] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 783.352813] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 783.353064] env[61991]: DEBUG nova.virt.hardware [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.359673] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfiguring VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 783.360305] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.360305] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Cleaning up deleted instances with incomplete migration {{(pid=61991) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 783.361835] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69ea89b8-1cf8-4626-aeec-496e4cc53af6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.377831] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.383605] env[61991]: DEBUG oslo_vmware.api [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 783.383605] env[61991]: value = "task-1129662" [ 783.383605] env[61991]: _type = "Task" [ 783.383605] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.393782] env[61991]: DEBUG oslo_vmware.api [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129662, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.536281] env[61991]: DEBUG nova.scheduler.client.report [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.744087] env[61991]: DEBUG nova.network.neutron [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.772894] env[61991]: DEBUG oslo_vmware.api [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129661, 'name': PowerOffVM_Task, 'duration_secs': 0.464154} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.773050] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.773223] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 783.773485] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd355645-8c63-42dc-915b-b1b672f909d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.817200] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129659, 'name': CreateVM_Task, 'duration_secs': 0.63083} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.817387] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 783.818507] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.818507] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.818793] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 783.819241] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5a9a081-3b62-4a52-9cf8-ed0f36334af1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.824179] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 783.824179] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5278fb3d-d4d1-2239-435c-01232da99baf" [ 783.824179] env[61991]: _type = "Task" [ 783.824179] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.828121] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.828370] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.828598] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.828785] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.829243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.831244] env[61991]: DEBUG nova.network.neutron [-] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.832659] env[61991]: INFO nova.compute.manager [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Terminating instance [ 783.837776] env[61991]: DEBUG nova.compute.manager [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.838011] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.839484] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e2df61-0718-496f-870c-a3f8ff981fda {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.847530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.847803] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Processing image e4b0de21-26e3-4d65-834b-7a893af37f7a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.848141] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.848385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.848582] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.848846] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 783.849047] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 783.849224] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleting the datastore file [datastore2] 0aa67bb1-59d3-42e4-8f12-47e61a130dfc {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 783.851842] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ab4ebb5-3675-480c-9bc8-2493f7ec380d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.854247] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef30866d-9e99-40e9-9bcb-538702770ccc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.856190] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.856673] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5669e0e2-c495-4d4d-8309-0f4c039301b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.863265] env[61991]: DEBUG oslo_vmware.api [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 783.863265] env[61991]: value = "task-1129665" [ 783.863265] env[61991]: _type = "Task" [ 783.863265] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.864715] env[61991]: DEBUG oslo_vmware.api [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 783.864715] env[61991]: value = "task-1129664" [ 783.864715] env[61991]: _type = "Task" [ 783.864715] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.869053] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.869274] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.874284] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-193ca6b4-6b00-4e2c-971e-dc87f8461050 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.885078] env[61991]: DEBUG oslo_vmware.api [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.893409] env[61991]: DEBUG oslo_vmware.api [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129664, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.893724] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 783.893724] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525ea5c8-1d80-0d52-60cb-bb34ae02cacc" [ 783.893724] env[61991]: _type = "Task" [ 783.893724] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.905068] env[61991]: DEBUG oslo_vmware.api [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.911190] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Preparing fetch location {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 783.911534] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Fetch image to [datastore1] OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e/OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e.vmdk {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 783.911730] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Downloading stream optimized image e4b0de21-26e3-4d65-834b-7a893af37f7a to [datastore1] OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e/OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e.vmdk on the data store datastore1 as vApp {{(pid=61991) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 783.912142] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Downloading image file data e4b0de21-26e3-4d65-834b-7a893af37f7a to the ESX as VM named 'OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e' {{(pid=61991) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 783.918036] env[61991]: DEBUG nova.network.neutron [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [{"id": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "address": "fa:16:3e:b6:b6:56", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.50", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap863ee2c0-0c", "ovs_interfaceid": "863ee2c0-0c5a-49c8-a443-3c1c3d73c458", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.995613] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 783.995613] env[61991]: value = "resgroup-9" [ 783.995613] env[61991]: _type = "ResourcePool" [ 783.995613] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 783.995980] env[61991]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-376034c6-ca32-4b6e-a905-700e4132f6f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.018074] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lease: (returnval){ [ 784.018074] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d1d89c-acce-d65d-1783-da9f2e844ecd" [ 784.018074] env[61991]: _type = "HttpNfcLease" [ 784.018074] env[61991]: } obtained for vApp import into resource pool (val){ [ 784.018074] env[61991]: value = "resgroup-9" [ 784.018074] env[61991]: _type = "ResourcePool" [ 784.018074] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 784.018383] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the lease: (returnval){ [ 784.018383] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d1d89c-acce-d65d-1783-da9f2e844ecd" [ 784.018383] env[61991]: _type = "HttpNfcLease" [ 784.018383] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 784.025681] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 784.025681] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d1d89c-acce-d65d-1783-da9f2e844ecd" [ 784.025681] env[61991]: _type = "HttpNfcLease" [ 784.025681] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 784.041031] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.041365] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.044228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.073s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.045773] env[61991]: INFO nova.compute.claims [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.247425] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.252787] env[61991]: DEBUG nova.compute.manager [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Received event network-changed-2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.253020] env[61991]: DEBUG nova.compute.manager [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Refreshing instance network info cache due to event network-changed-2d37704c-b54b-459f-b593-12a9062c3601. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.253262] env[61991]: DEBUG oslo_concurrency.lockutils [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] Acquiring lock "refresh_cache-e3e5ec72-eec1-4155-bab2-038e3d5e38a8" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.253413] env[61991]: DEBUG oslo_concurrency.lockutils [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] Acquired lock "refresh_cache-e3e5ec72-eec1-4155-bab2-038e3d5e38a8" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.253600] env[61991]: DEBUG nova.network.neutron [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Refreshing network info cache for port 2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.338787] env[61991]: INFO nova.compute.manager [-] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Took 1.36 seconds to deallocate network for instance. [ 784.383585] env[61991]: DEBUG oslo_vmware.api [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129665, 'name': PowerOffVM_Task, 'duration_secs': 0.3181} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.387748] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.388043] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 784.388926] env[61991]: DEBUG oslo_vmware.api [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129664, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309251} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.389727] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6f8e0f5-b8a8-409d-baf5-a94d45b72743 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.394599] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.394782] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.394965] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.395205] env[61991]: INFO nova.compute.manager [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Took 1.15 seconds to destroy the instance on the hypervisor. [ 784.395475] env[61991]: DEBUG oslo.service.loopingcall [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.397048] env[61991]: DEBUG nova.compute.manager [-] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.397172] env[61991]: DEBUG nova.network.neutron [-] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 784.400535] env[61991]: DEBUG nova.compute.manager [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-changed-a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.400653] env[61991]: DEBUG nova.compute.manager [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing instance network info cache due to event network-changed-a20e94c9-8ee3-4696-b020-423d80f00dc9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.400921] env[61991]: DEBUG oslo_concurrency.lockutils [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.401131] env[61991]: DEBUG oslo_concurrency.lockutils [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.401269] env[61991]: DEBUG nova.network.neutron [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing network info cache for port a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.409262] env[61991]: DEBUG oslo_vmware.api [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129662, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.424068] env[61991]: DEBUG oslo_concurrency.lockutils [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Releasing lock "refresh_cache-c14d99a3-950d-44d0-b330-3031a1f8a2be" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.424068] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 784.424068] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d7bd62e-badb-4d7f-9a59-61b345ab30c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.430760] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 784.430760] env[61991]: value = "task-1129668" [ 784.430760] env[61991]: _type = "Task" [ 784.430760] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.443230] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.527061] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 784.527061] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d1d89c-acce-d65d-1783-da9f2e844ecd" [ 784.527061] env[61991]: _type = "HttpNfcLease" [ 784.527061] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 784.551206] env[61991]: DEBUG nova.compute.utils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.557946] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 784.558231] env[61991]: DEBUG nova.network.neutron [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 784.634389] env[61991]: DEBUG nova.policy [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc2e1f75d6bd44df96ad5b229d883b31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25647cd0715f490092623f4ead833738', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.846223] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.887309] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 784.887530] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 784.887770] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleting the datastore file [datastore2] c279a82f-bf4c-4a59-b04a-38d31070b0e5 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.888343] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7fe0602-7191-4ca3-a20c-852884d214e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.898555] env[61991]: DEBUG oslo_vmware.api [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129662, 'name': ReconfigVM_Task, 'duration_secs': 1.074351} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.901996] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.903244] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfigured VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 784.904949] env[61991]: DEBUG oslo_vmware.api [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for the task: (returnval){ [ 784.904949] env[61991]: value = "task-1129669" [ 784.904949] env[61991]: _type = "Task" [ 784.904949] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.916771] env[61991]: DEBUG oslo_vmware.api [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.940250] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129668, 'name': PowerOffVM_Task, 'duration_secs': 0.31646} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.940666] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.941347] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:18:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b41f5b31-f99f-43b3-b2c2-a0cd889c0027',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-424611864',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.941641] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.941850] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.945019] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.948499] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4222a918-da5d-482a-8e9e-2654737adce8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.964666] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 784.964666] env[61991]: value = "task-1129670" [ 784.964666] env[61991]: _type = "Task" [ 784.964666] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.976774] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129670, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.026876] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 785.026876] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d1d89c-acce-d65d-1783-da9f2e844ecd" [ 785.026876] env[61991]: _type = "HttpNfcLease" [ 785.026876] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 785.027231] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 785.027231] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d1d89c-acce-d65d-1783-da9f2e844ecd" [ 785.027231] env[61991]: _type = "HttpNfcLease" [ 785.027231] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 785.027988] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e252dadd-d0ea-4215-acb0-c980fecfc37f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.035299] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52603895-8fd6-15b9-d2ee-a275b671e7e6/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 785.035773] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52603895-8fd6-15b9-d2ee-a275b671e7e6/disk-0.vmdk. {{(pid=61991) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 785.100919] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.119246] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.122616] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb6b181-6211-457e-9de2-dcc6a1894dc0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.130696] env[61991]: DEBUG nova.network.neutron [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Successfully created port: 3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.135725] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3d8a718c-3c58-4ba0-8dfe-432fc7fda02a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.140807] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 785.141708] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ddbbdf0e-9ca3-4572-9b37-aefea9d88374 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.220992] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 785.221263] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 785.221476] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleting the datastore file [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.221808] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2955025-c720-48c9-9b4d-d44efb9ba3c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.227168] env[61991]: DEBUG nova.network.neutron [-] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.231489] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 785.231489] env[61991]: value = "task-1129672" [ 785.231489] env[61991]: _type = "Task" [ 785.231489] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.240549] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.251411] env[61991]: DEBUG nova.network.neutron [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Updated VIF entry in instance network info cache for port 2d37704c-b54b-459f-b593-12a9062c3601. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.251805] env[61991]: DEBUG nova.network.neutron [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Updating instance_info_cache with network_info: [{"id": "2d37704c-b54b-459f-b593-12a9062c3601", "address": "fa:16:3e:a3:e5:75", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d37704c-b5", "ovs_interfaceid": "2d37704c-b54b-459f-b593-12a9062c3601", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.407134] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1420b319-b092-448f-ad5d-7f480a1a1726 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.449s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.409074] env[61991]: DEBUG nova.network.neutron [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updated VIF entry in instance network info cache for port a20e94c9-8ee3-4696-b020-423d80f00dc9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.409308] env[61991]: DEBUG nova.network.neutron [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "address": "fa:16:3e:5c:e6:a5", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20e94c9-8e", "ovs_interfaceid": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.425459] env[61991]: DEBUG oslo_vmware.api [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Task: {'id': task-1129669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242927} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.428056] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.428345] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 785.428475] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.428661] env[61991]: INFO nova.compute.manager [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Took 1.59 seconds to destroy the instance on the hypervisor. [ 785.428948] env[61991]: DEBUG oslo.service.loopingcall [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.429883] env[61991]: DEBUG nova.compute.manager [-] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.430909] env[61991]: DEBUG nova.network.neutron [-] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 785.476880] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129670, 'name': ReconfigVM_Task, 'duration_secs': 0.165924} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.480472] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31278bf-f75c-45db-ae1e-a0da667b75b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.502421] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:18:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b41f5b31-f99f-43b3-b2c2-a0cd889c0027',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-424611864',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 785.502733] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 785.502902] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.503146] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 785.503305] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.503456] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 785.503665] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 785.503830] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 785.503995] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 785.504194] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 785.504369] env[61991]: DEBUG nova.virt.hardware [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.511907] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-389e8052-c653-4885-b67d-ba2e9fe06159 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.519639] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 785.519639] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52039119-7749-f1f1-0f8b-90a3854cc34a" [ 785.519639] env[61991]: _type = "Task" [ 785.519639] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.532726] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52039119-7749-f1f1-0f8b-90a3854cc34a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.555320] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f7eea8-ae4f-4bd2-9bc1-8ba9a8ef180f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.565070] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48def05a-013e-4964-9d46-5719e67f30c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.605051] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cf1f85-2169-4aff-a25b-1b54ea50162a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.624518] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa6cc9c-a4e9-4583-a2b4-3327d050b0e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.648466] env[61991]: DEBUG nova.compute.provider_tree [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.730092] env[61991]: INFO nova.compute.manager [-] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Took 1.33 seconds to deallocate network for instance. [ 785.755032] env[61991]: DEBUG oslo_concurrency.lockutils [req-ae2dc381-db79-464e-a0a8-8af60b3ea277 req-bf02ce08-856c-403f-9e08-f45a8ac9cd0e service nova] Releasing lock "refresh_cache-e3e5ec72-eec1-4155-bab2-038e3d5e38a8" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.755518] env[61991]: DEBUG oslo_vmware.api [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322339} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.757826] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.758029] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 785.758293] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.800956] env[61991]: INFO nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleted allocations for instance f78ef63b-453e-45d3-959b-4b0c1922b53e [ 785.920202] env[61991]: DEBUG oslo_concurrency.lockutils [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.920932] env[61991]: DEBUG nova.compute.manager [req-a0a12e7a-494d-4c28-bf0d-9d7cc3cbd5b0 req-36858dc7-9611-4406-bef4-44db36fe4134 service nova] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Received event network-vif-deleted-179a25d5-af58-45b5-b510-6f9064bddb3c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.977870] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Completed reading data from the image iterator. {{(pid=61991) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 785.978148] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52603895-8fd6-15b9-d2ee-a275b671e7e6/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 785.979492] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a00bdc-f48d-464e-8a64-97ed53731f3f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.986559] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52603895-8fd6-15b9-d2ee-a275b671e7e6/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 785.987202] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52603895-8fd6-15b9-d2ee-a275b671e7e6/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 785.987202] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-965bde56-bcb4-4956-8312-97e3e389f295 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.031674] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52039119-7749-f1f1-0f8b-90a3854cc34a, 'name': SearchDatastore_Task, 'duration_secs': 0.010289} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.036841] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 786.037228] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f61e7bf-1bb1-409f-a282-baf3ab5af414 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.058692] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 786.058692] env[61991]: value = "task-1129673" [ 786.058692] env[61991]: _type = "Task" [ 786.058692] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.066912] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.115756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.115756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.115756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.115756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.115756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.118826] env[61991]: INFO nova.compute.manager [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Terminating instance [ 786.122174] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.125254] env[61991]: DEBUG nova.compute.manager [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 786.125254] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.127751] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17499a1f-cdb9-4b54-8f6c-40a49d76eed1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.134695] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.134984] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23680a40-a838-425e-ae2e-a1ffb61ccf73 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.141732] env[61991]: DEBUG oslo_vmware.api [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 786.141732] env[61991]: value = "task-1129674" [ 786.141732] env[61991]: _type = "Task" [ 786.141732] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.149982] env[61991]: DEBUG oslo_vmware.api [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.152598] env[61991]: DEBUG nova.scheduler.client.report [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.166287] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.166557] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.166773] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.166970] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.168480] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.168480] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.168480] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.168480] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.168480] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.168480] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.168715] env[61991]: DEBUG nova.virt.hardware [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.171502] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ade0f0-2c35-4772-be2f-0d7ed2402188 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.186680] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3c3385-82a2-4472-ad7a-bf5f1f46bf6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.219018] env[61991]: DEBUG oslo_vmware.rw_handles [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52603895-8fd6-15b9-d2ee-a275b671e7e6/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 786.219285] env[61991]: INFO nova.virt.vmwareapi.images [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Downloaded image file data e4b0de21-26e3-4d65-834b-7a893af37f7a [ 786.220174] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c37054b-b631-4f59-a698-ca94f41ca812 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.228032] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9855-8428-2f12-1193-c18147daeb18/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 786.228032] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0141417-1aaf-40fa-a475-f765db08702c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.246444] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-644296a2-38c8-4ed2-be02-8671e9d957b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.250418] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.250736] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9855-8428-2f12-1193-c18147daeb18/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 786.250920] env[61991]: ERROR oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9855-8428-2f12-1193-c18147daeb18/disk-0.vmdk due to incomplete transfer. [ 786.251549] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-aef18434-30e8-4846-9b36-fe7403f485cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.262214] env[61991]: DEBUG oslo_vmware.rw_handles [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9855-8428-2f12-1193-c18147daeb18/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 786.262642] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Uploaded image 5b317551-bad7-4685-b760-e9c291dff08c to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 786.264195] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 786.264474] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f20498c0-913d-4bb7-bd08-b4a9b9d97385 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.270531] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 786.270531] env[61991]: value = "task-1129676" [ 786.270531] env[61991]: _type = "Task" [ 786.270531] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.279861] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129676, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.306607] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.320485] env[61991]: INFO nova.virt.vmwareapi.images [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] The imported VM was unregistered [ 786.323284] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Caching image {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 786.323600] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.323898] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7c3e398-d4bd-4d59-af5f-e3de3aa5033f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.326510] env[61991]: DEBUG nova.network.neutron [-] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.338699] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Created directory with path [datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.338856] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e/OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e.vmdk to [datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk. {{(pid=61991) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 786.338981] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7432ed32-cb5f-4b82-b00a-db28afcffd13 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.346372] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 786.346372] env[61991]: value = "task-1129677" [ 786.346372] env[61991]: _type = "Task" [ 786.346372] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.354472] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129677, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.578141] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129673, 'name': ReconfigVM_Task, 'duration_secs': 0.252426} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.578840] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 786.579379] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b748cd-3c95-4a6a-9c27-4ff2ffb22928 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.603815] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.604649] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63b3118a-605a-435d-8b30-42d150437468 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.627414] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 786.627414] env[61991]: value = "task-1129678" [ 786.627414] env[61991]: _type = "Task" [ 786.627414] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.639161] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129678, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.647577] env[61991]: DEBUG nova.compute.manager [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-vif-unplugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.647810] env[61991]: DEBUG oslo_concurrency.lockutils [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.648587] env[61991]: DEBUG oslo_concurrency.lockutils [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.648587] env[61991]: DEBUG oslo_concurrency.lockutils [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.648587] env[61991]: DEBUG nova.compute.manager [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] No waiting events found dispatching network-vif-unplugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 786.648782] env[61991]: WARNING nova.compute.manager [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received unexpected event network-vif-unplugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 for instance with vm_state shelved_offloaded and task_state None. [ 786.648782] env[61991]: DEBUG nova.compute.manager [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.648909] env[61991]: DEBUG nova.compute.manager [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing instance network info cache due to event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 786.649119] env[61991]: DEBUG oslo_concurrency.lockutils [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.649245] env[61991]: DEBUG oslo_concurrency.lockutils [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.649397] env[61991]: DEBUG nova.network.neutron [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 786.661434] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.662446] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.666691] env[61991]: DEBUG oslo_vmware.api [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129674, 'name': PowerOffVM_Task, 'duration_secs': 0.215971} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.667390] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.502s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.668238] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.670043] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.301s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.670242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.672512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.816s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.672917] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.676139] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.754s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.678323] env[61991]: INFO nova.compute.claims [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.682686] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.682948] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.689628] env[61991]: DEBUG nova.compute.manager [req-442193b8-56ec-4dd3-b2b9-359032a7d8d0 req-9a0562e7-6b34-41e2-9f5a-e8eb0c6d2e2c service nova] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Received event network-vif-deleted-6a28231d-a28f-49ea-8339-3ca4a02684bb {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.689628] env[61991]: DEBUG nova.compute.manager [req-442193b8-56ec-4dd3-b2b9-359032a7d8d0 req-9a0562e7-6b34-41e2-9f5a-e8eb0c6d2e2c service nova] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Received event network-vif-deleted-2d6d64e8-636f-4478-93ab-f1ca63a74ce1 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.689628] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e3f6b8b-e1d7-4250-ad28-79aa97d24c29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.721945] env[61991]: INFO nova.scheduler.client.report [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Deleted allocations for instance 3e06e424-b64e-4fd2-8013-27760200c41d [ 786.732088] env[61991]: INFO nova.scheduler.client.report [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Deleted allocations for instance 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349 [ 786.760159] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.760379] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.760566] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Deleting the datastore file [datastore2] 700ce4a7-d597-449a-9379-0cfb0c8f82fc {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.761138] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-001ad82f-786a-4579-a54a-3a12191d1d44 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.769093] env[61991]: DEBUG oslo_vmware.api [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for the task: (returnval){ [ 786.769093] env[61991]: value = "task-1129680" [ 786.769093] env[61991]: _type = "Task" [ 786.769093] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.791999] env[61991]: DEBUG oslo_vmware.api [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.791999] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129676, 'name': Destroy_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.829364] env[61991]: INFO nova.compute.manager [-] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Took 1.40 seconds to deallocate network for instance. [ 786.859965] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129677, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.893785] env[61991]: DEBUG nova.network.neutron [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Successfully updated port: 3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.136626] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129678, 'name': ReconfigVM_Task, 'duration_secs': 0.504597} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.136947] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Reconfigured VM instance instance-0000002b to attach disk [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be/c14d99a3-950d-44d0-b330-3031a1f8a2be.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.137841] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864bc95d-bc68-40d3-a51d-b9fd1e1232dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.158810] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d46a2c-2c5c-4be8-979a-1da4cc7ab2bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.178044] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f1a465-fd5b-4d6c-a22c-e613d48a2ac4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.196447] env[61991]: DEBUG nova.compute.utils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.203775] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.204049] env[61991]: DEBUG nova.network.neutron [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 787.206789] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95511bb7-46b0-4aec-8162-ec852caecd88 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.213045] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e850598a-649b-4692-bc1e-80bad001683a tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 47.044s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.214790] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 23.874s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.215071] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "78631e83-e574-4a90-b062-176b2d12d681-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.215323] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.215513] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.220196] env[61991]: INFO nova.compute.manager [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Terminating instance [ 787.225357] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.226094] env[61991]: DEBUG nova.compute.manager [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 787.226308] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.226575] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f1d9515-c7b5-4802-93ca-14c5b0dca49c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.228941] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee8acb7c-7ef8-474a-8c32-dd5ed5f13e72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.235429] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e394e71d-cd76-42ba-ba72-c8a5e3811a0d tempest-SecurityGroupsTestJSON-653957827 tempest-SecurityGroupsTestJSON-653957827-project-member] Lock "3e06e424-b64e-4fd2-8013-27760200c41d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.551s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.240582] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 787.240582] env[61991]: value = "task-1129681" [ 787.240582] env[61991]: _type = "Task" [ 787.240582] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.241043] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c3a0f6ad-fd38-4f2f-ba54-b94e25d29b0e tempest-VolumesAdminNegativeTest-180125387 tempest-VolumesAdminNegativeTest-180125387-project-member] Lock "4b3c05a3-c40a-40c8-8501-5c1b8e4ea349" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.794s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.251178] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.258656] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2898dbc-ca00-4ed7-874f-f31b3ca9ba31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.280275] env[61991]: DEBUG oslo_vmware.api [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Task: {'id': task-1129680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285632} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.283721] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.287101] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.287101] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.287101] env[61991]: INFO nova.compute.manager [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Took 1.16 seconds to destroy the instance on the hypervisor. [ 787.287101] env[61991]: DEBUG oslo.service.loopingcall [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.294827] env[61991]: DEBUG nova.policy [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2d8e09acc6a41e29af8b490b055dc30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5322f4e5522f4a3d8f915d056e11ebce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.296536] env[61991]: DEBUG nova.compute.manager [-] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.296653] env[61991]: DEBUG nova.network.neutron [-] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.298747] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 78631e83-e574-4a90-b062-176b2d12d681 could not be found. [ 787.300211] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.300211] env[61991]: INFO nova.compute.manager [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Took 0.07 seconds to destroy the instance on the hypervisor. [ 787.300211] env[61991]: DEBUG oslo.service.loopingcall [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.302347] env[61991]: DEBUG nova.compute.manager [-] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.302458] env[61991]: DEBUG nova.network.neutron [-] [instance: 78631e83-e574-4a90-b062-176b2d12d681] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.308941] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129676, 'name': Destroy_Task, 'duration_secs': 1.01744} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.308941] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroyed the VM [ 787.309063] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 787.309283] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c61c57b7-6ba0-4dba-ab92-65b172bffd1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.315412] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 787.315412] env[61991]: value = "task-1129682" [ 787.315412] env[61991]: _type = "Task" [ 787.315412] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.323824] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129682, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.338029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.359472] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129677, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.398503] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "refresh_cache-3e62f219-7fec-470f-9d0a-e292d62e076c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.398503] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "refresh_cache-3e62f219-7fec-470f-9d0a-e292d62e076c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.398503] env[61991]: DEBUG nova.network.neutron [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 787.467469] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-e08c38be-49fa-4a29-9cb1-37e15930e258" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.467753] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-e08c38be-49fa-4a29-9cb1-37e15930e258" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.468147] env[61991]: DEBUG nova.objects.instance [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'flavor' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 787.515929] env[61991]: DEBUG nova.network.neutron [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updated VIF entry in instance network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 787.516490] env[61991]: DEBUG nova.network.neutron [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4b255035-eb", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.707319] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.757149] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.831993] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129682, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.865238] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129677, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.930759] env[61991]: DEBUG nova.network.neutron [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.992728] env[61991]: DEBUG nova.network.neutron [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Successfully created port: 50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.021719] env[61991]: DEBUG oslo_concurrency.lockutils [req-dab41bb1-3e23-44c1-b659-8634c59f009a req-3b0a3a92-f949-46e6-a2e6-ca318d0bdf3d service nova] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.050674] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef5fbce-8d6b-4e24-9a2c-97a87d58361c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.059745] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def4454f-f391-48df-a806-190006edfc24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.093490] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04cc88b-4cc4-4e91-a3ea-c67deff16258 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.104385] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1d167e-3f52-45b8-8dbc-6dad05fcab12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.118625] env[61991]: DEBUG nova.network.neutron [-] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.120236] env[61991]: DEBUG nova.compute.provider_tree [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 788.160997] env[61991]: DEBUG nova.network.neutron [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Updating instance_info_cache with network_info: [{"id": "3ea67247-a11c-4dfb-8875-ffc387e7678e", "address": "fa:16:3e:bd:a4:ce", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea67247-a1", "ovs_interfaceid": "3ea67247-a11c-4dfb-8875-ffc387e7678e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.163478] env[61991]: DEBUG nova.objects.instance [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'pci_requests' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.263134] env[61991]: DEBUG oslo_vmware.api [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129681, 'name': PowerOnVM_Task, 'duration_secs': 0.640274} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.263590] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.328818] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129682, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.363763] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129677, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.622543] env[61991]: INFO nova.compute.manager [-] [instance: 78631e83-e574-4a90-b062-176b2d12d681] Took 1.32 seconds to deallocate network for instance. [ 788.644494] env[61991]: ERROR nova.scheduler.client.report [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [req-b9890fee-9a92-43d2-95d9-6e3ad94de3c9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b9890fee-9a92-43d2-95d9-6e3ad94de3c9"}]} [ 788.663033] env[61991]: DEBUG nova.scheduler.client.report [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 788.665759] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "refresh_cache-3e62f219-7fec-470f-9d0a-e292d62e076c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.666231] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Instance network_info: |[{"id": "3ea67247-a11c-4dfb-8875-ffc387e7678e", "address": "fa:16:3e:bd:a4:ce", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea67247-a1", "ovs_interfaceid": "3ea67247-a11c-4dfb-8875-ffc387e7678e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 788.666747] env[61991]: DEBUG nova.objects.base [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Object Instance<80d05278-b516-4408-94b0-11bc93500b5c> lazy-loaded attributes: flavor,pci_requests {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 788.667011] env[61991]: DEBUG nova.network.neutron [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 788.670618] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:a4:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc845e3-654b-43c6-acea-dde1084f0ad0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ea67247-a11c-4dfb-8875-ffc387e7678e', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.680822] env[61991]: DEBUG oslo.service.loopingcall [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.684051] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.684051] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b7466e5-b837-48ef-983e-cbb31686f8a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.699554] env[61991]: DEBUG nova.scheduler.client.report [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 788.699780] env[61991]: DEBUG nova.compute.provider_tree [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 788.707557] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.707557] env[61991]: value = "task-1129683" [ 788.707557] env[61991]: _type = "Task" [ 788.707557] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.712589] env[61991]: DEBUG nova.scheduler.client.report [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 788.718342] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129683, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.723690] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.739502] env[61991]: DEBUG nova.scheduler.client.report [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 788.749783] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.750057] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.750230] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.750458] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.750618] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.750770] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.750981] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.751162] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.751334] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.751499] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.751681] env[61991]: DEBUG nova.virt.hardware [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.752592] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7708a17-a25c-40fb-a3fc-08adcbe49367 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.760365] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd562aa-c2da-4cec-93a7-76f70c4ccaa7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.829198] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129682, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.836592] env[61991]: DEBUG nova.policy [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.864292] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129677, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.420742} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.865714] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e/OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e.vmdk to [datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk. [ 788.865924] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Cleaning up location [datastore1] OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 788.866108] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_4f332539-046b-471a-a010-83079092b37e {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.866539] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b1d8894-63fa-4937-9396-a61f2281a715 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.876817] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 788.876817] env[61991]: value = "task-1129684" [ 788.876817] env[61991]: _type = "Task" [ 788.876817] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.884042] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.056218] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38714a1-8224-477b-8174-4ce20f3fd5a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.066543] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540dbf65-ecf7-4a40-b68b-a1acb880c6eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.096316] env[61991]: DEBUG nova.network.neutron [-] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.097929] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f94bd33-4801-47ae-92ae-8b00e1ae5c33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.106234] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f5debc-49f6-46e4-84dc-659e5dbb0e65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.120828] env[61991]: DEBUG nova.compute.provider_tree [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.218443] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129683, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.287023] env[61991]: INFO nova.compute.manager [None req-41c29d52-3a8f-48db-9a27-d169e4300363 tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance to original state: 'active' [ 789.328481] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129682, 'name': RemoveSnapshot_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.383127] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096212} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.383419] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.383630] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.383947] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk to [datastore1] e3e5ec72-eec1-4155-bab2-038e3d5e38a8/e3e5ec72-eec1-4155-bab2-038e3d5e38a8.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.384258] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0267d013-8bbf-4e3f-a3a9-b1dae075d84c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.390473] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 789.390473] env[61991]: value = "task-1129685" [ 789.390473] env[61991]: _type = "Task" [ 789.390473] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.399963] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.601942] env[61991]: INFO nova.compute.manager [-] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Took 2.31 seconds to deallocate network for instance. [ 789.651857] env[61991]: DEBUG nova.scheduler.client.report [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 789.652173] env[61991]: DEBUG nova.compute.provider_tree [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 83 to 84 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 789.652367] env[61991]: DEBUG nova.compute.provider_tree [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.658933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d519ba80-0ad7-4e66-9821-3593981a10da tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "78631e83-e574-4a90-b062-176b2d12d681" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.444s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.718858] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129683, 'name': CreateVM_Task, 'duration_secs': 0.971509} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.719185] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.719763] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.719934] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.720276] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.720529] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7a30665-7e45-470a-baf2-db8ac20ec113 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.725316] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 789.725316] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ebe32e-2f98-ab78-c911-f857128e1375" [ 789.725316] env[61991]: _type = "Task" [ 789.725316] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.732626] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ebe32e-2f98-ab78-c911-f857128e1375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.829872] env[61991]: DEBUG oslo_vmware.api [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129682, 'name': RemoveSnapshot_Task, 'duration_secs': 2.22841} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.830317] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 789.830663] env[61991]: INFO nova.compute.manager [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 18.54 seconds to snapshot the instance on the hypervisor. [ 789.900883] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.109218] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.161099] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.485s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.161884] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.165371] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.542s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.166958] env[61991]: INFO nova.compute.claims [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.215094] env[61991]: DEBUG nova.network.neutron [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Successfully updated port: 50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.237739] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ebe32e-2f98-ab78-c911-f857128e1375, 'name': SearchDatastore_Task, 'duration_secs': 0.016731} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.238628] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.239193] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.240017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.240017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.240849] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.241290] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79fd1168-c4bb-404e-80c6-5f1466d6492a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.260047] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.260047] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.260047] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6220295a-8ab0-4898-ad65-4c1a9cc91242 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.266639] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 790.266639] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b4a77-638d-8633-8f9a-c2a093f286aa" [ 790.266639] env[61991]: _type = "Task" [ 790.266639] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.276426] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b4a77-638d-8633-8f9a-c2a093f286aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.392409] env[61991]: DEBUG nova.compute.manager [None req-49382856-7d41-4578-974e-a82f8c6574e9 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Found 1 images (rotation: 2) {{(pid=61991) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 790.403667] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.561264] env[61991]: DEBUG nova.compute.manager [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Received event network-vif-plugged-3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.561405] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] Acquiring lock "3e62f219-7fec-470f-9d0a-e292d62e076c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.561816] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.562122] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.562330] env[61991]: DEBUG nova.compute.manager [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] No waiting events found dispatching network-vif-plugged-3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.562574] env[61991]: WARNING nova.compute.manager [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Received unexpected event network-vif-plugged-3ea67247-a11c-4dfb-8875-ffc387e7678e for instance with vm_state building and task_state spawning. [ 790.562859] env[61991]: DEBUG nova.compute.manager [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Received event network-changed-3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.563044] env[61991]: DEBUG nova.compute.manager [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Refreshing instance network info cache due to event network-changed-3ea67247-a11c-4dfb-8875-ffc387e7678e. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.563244] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] Acquiring lock "refresh_cache-3e62f219-7fec-470f-9d0a-e292d62e076c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.563388] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] Acquired lock "refresh_cache-3e62f219-7fec-470f-9d0a-e292d62e076c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.563549] env[61991]: DEBUG nova.network.neutron [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Refreshing network info cache for port 3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.575687] env[61991]: DEBUG nova.network.neutron [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Successfully updated port: e08c38be-49fa-4a29-9cb1-37e15930e258 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.584063] env[61991]: DEBUG nova.compute.manager [req-7b0dc42e-e063-4196-95ed-ace4ec9bd010 req-e283660b-b03d-4019-9a4f-3b1172043b3a service nova] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Received event network-vif-deleted-2e472308-0d44-4e9e-ae6e-3afaad4796b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.675563] env[61991]: DEBUG nova.compute.utils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.676484] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.676658] env[61991]: DEBUG nova.network.neutron [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.717955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "refresh_cache-884e465b-e14f-4114-81a6-40e2e8694558" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.718159] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquired lock "refresh_cache-884e465b-e14f-4114-81a6-40e2e8694558" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.718347] env[61991]: DEBUG nova.network.neutron [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 790.750709] env[61991]: DEBUG nova.policy [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af6cfe94f59140f1b85bf2a1e8593e70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79c013b6c254403b91f520e2af18027f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.777707] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b4a77-638d-8633-8f9a-c2a093f286aa, 'name': SearchDatastore_Task, 'duration_secs': 0.078239} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.778667] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7046ddef-6edf-4fc5-952e-f503eaf5b650 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.784767] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 790.784767] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f8701d-3b53-ffb7-a325-5b10b42355c9" [ 790.784767] env[61991]: _type = "Task" [ 790.784767] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.793589] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f8701d-3b53-ffb7-a325-5b10b42355c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.905865] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.079633] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.079633] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.079800] env[61991]: DEBUG nova.network.neutron [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.180225] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.265618] env[61991]: DEBUG nova.network.neutron [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.298721] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f8701d-3b53-ffb7-a325-5b10b42355c9, 'name': SearchDatastore_Task, 'duration_secs': 0.085694} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.301201] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.302110] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3e62f219-7fec-470f-9d0a-e292d62e076c/3e62f219-7fec-470f-9d0a-e292d62e076c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.302110] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c9a899a-b20b-4b9d-a31d-b003c4c6b2ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.314227] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 791.314227] env[61991]: value = "task-1129686" [ 791.314227] env[61991]: _type = "Task" [ 791.314227] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.322316] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.323284] env[61991]: DEBUG nova.network.neutron [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Successfully created port: c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.408540] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.531680] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce80e75a-6a56-4ee2-8ec5-4a7dedc1f469 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.536375] env[61991]: DEBUG nova.network.neutron [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Updating instance_info_cache with network_info: [{"id": "50b6a76d-9313-4b7c-b577-09c086277647", "address": "fa:16:3e:35:98:fa", "network": {"id": "e3aa1719-f3cd-4ad6-bd6d-5f950bb7d8f2", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1572203683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5322f4e5522f4a3d8f915d056e11ebce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50b6a76d-93", "ovs_interfaceid": "50b6a76d-9313-4b7c-b577-09c086277647", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.546544] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8912ad-dcb3-4866-aef5-7969eb81a0fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.583979] env[61991]: DEBUG nova.network.neutron [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Updated VIF entry in instance network info cache for port 3ea67247-a11c-4dfb-8875-ffc387e7678e. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.583979] env[61991]: DEBUG nova.network.neutron [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Updating instance_info_cache with network_info: [{"id": "3ea67247-a11c-4dfb-8875-ffc387e7678e", "address": "fa:16:3e:bd:a4:ce", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ea67247-a1", "ovs_interfaceid": "3ea67247-a11c-4dfb-8875-ffc387e7678e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.584503] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f249956a-2910-47dd-aff0-8b5976a6184d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.595442] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58925f45-3361-4836-ab81-e01ed755203f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.612841] env[61991]: DEBUG nova.compute.provider_tree [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.652321] env[61991]: WARNING nova.network.neutron [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] e6f9d471-4f42-4781-93c3-f4a9ff84317b already exists in list: networks containing: ['e6f9d471-4f42-4781-93c3-f4a9ff84317b']. ignoring it [ 791.652574] env[61991]: WARNING nova.network.neutron [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] e6f9d471-4f42-4781-93c3-f4a9ff84317b already exists in list: networks containing: ['e6f9d471-4f42-4781-93c3-f4a9ff84317b']. ignoring it [ 791.824684] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.906431] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.038651] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Releasing lock "refresh_cache-884e465b-e14f-4114-81a6-40e2e8694558" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.038975] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Instance network_info: |[{"id": "50b6a76d-9313-4b7c-b577-09c086277647", "address": "fa:16:3e:35:98:fa", "network": {"id": "e3aa1719-f3cd-4ad6-bd6d-5f950bb7d8f2", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1572203683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5322f4e5522f4a3d8f915d056e11ebce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50b6a76d-93", "ovs_interfaceid": "50b6a76d-9313-4b7c-b577-09c086277647", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.039789] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:98:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '50b6a76d-9313-4b7c-b577-09c086277647', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.050368] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Creating folder: Project (5322f4e5522f4a3d8f915d056e11ebce). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.050718] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5fa37ab-7b35-4d09-be6e-04dd2deda009 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.062909] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Created folder: Project (5322f4e5522f4a3d8f915d056e11ebce) in parent group-v246753. [ 792.063138] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Creating folder: Instances. Parent ref: group-v246864. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.063434] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dafb3c37-8632-4a3b-9c01-e6651cc3095c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.072653] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Created folder: Instances in parent group-v246864. [ 792.072917] env[61991]: DEBUG oslo.service.loopingcall [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.073157] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.073384] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6336d3cc-1528-4d97-9c6e-ffbedd7a501b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.093471] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba1d3ebd-a97c-479c-b708-a7b347d5fad8 req-c8298485-5467-4afd-ac54-13547db49c17 service nova] Releasing lock "refresh_cache-3e62f219-7fec-470f-9d0a-e292d62e076c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.099487] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.099487] env[61991]: value = "task-1129689" [ 792.099487] env[61991]: _type = "Task" [ 792.099487] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.107512] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129689, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.116270] env[61991]: DEBUG nova.scheduler.client.report [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.191547] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.225414] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.225414] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.225414] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.225414] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.225832] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.225832] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.226033] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.227777] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.228008] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.228570] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.228570] env[61991]: DEBUG nova.virt.hardware [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.229951] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d96380-20e6-4a6c-89cf-deae84010892 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.240583] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1bdd8f-a44e-4c57-a067-b502ad44ba3c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.299879] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "15215bba-8160-408b-9e8f-afbb9e71d50e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.301312] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.331691] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.998458} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.332470] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3e62f219-7fec-470f-9d0a-e292d62e076c/3e62f219-7fec-470f-9d0a-e292d62e076c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.332779] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.333127] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b7bae5f-cb7b-4f69-8fcb-9bdd1aed6b7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.340713] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 792.340713] env[61991]: value = "task-1129690" [ 792.340713] env[61991]: _type = "Task" [ 792.340713] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.351835] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.370366] env[61991]: DEBUG nova.network.neutron [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "address": "fa:16:3e:5c:e6:a5", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20e94c9-8e", "ovs_interfaceid": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e08c38be-49fa-4a29-9cb1-37e15930e258", "address": "fa:16:3e:41:e8:b9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape08c38be-49", "ovs_interfaceid": "e08c38be-49fa-4a29-9cb1-37e15930e258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.407523] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129685, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.525231} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.407878] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e4b0de21-26e3-4d65-834b-7a893af37f7a/e4b0de21-26e3-4d65-834b-7a893af37f7a.vmdk to [datastore1] e3e5ec72-eec1-4155-bab2-038e3d5e38a8/e3e5ec72-eec1-4155-bab2-038e3d5e38a8.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.408726] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7238e491-20ec-40c2-b542-4d14fb0058e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.434055] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] e3e5ec72-eec1-4155-bab2-038e3d5e38a8/e3e5ec72-eec1-4155-bab2-038e3d5e38a8.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.435032] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd5243d7-f4b9-4fa8-a09e-5ba2f68c8541 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.455484] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 792.455484] env[61991]: value = "task-1129691" [ 792.455484] env[61991]: _type = "Task" [ 792.455484] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.466703] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129691, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.609837] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129689, 'name': CreateVM_Task, 'duration_secs': 0.428179} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.610166] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 792.611081] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.611375] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.611799] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 792.612184] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d07967ba-bf07-4e31-a83e-025a8d8fb7cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.617053] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 792.617053] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529db4da-4b23-ce9f-0885-c20b0b63d417" [ 792.617053] env[61991]: _type = "Task" [ 792.617053] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.621818] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.622818] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 792.627069] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.147s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.627069] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.630702] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.312s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.631377] env[61991]: INFO nova.compute.claims [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.643599] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529db4da-4b23-ce9f-0885-c20b0b63d417, 'name': SearchDatastore_Task, 'duration_secs': 0.01106} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.645250] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.645250] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 792.645250] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.645250] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.645250] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 792.645595] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-825b921f-aa77-44f7-b798-cc1928f5b4d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.655456] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 792.655456] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 792.656351] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c22a2d2-b572-4818-bddb-565217c299a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.668454] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 792.668454] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527e6ea2-386d-00d8-4eb8-14f1f5c9ddb1" [ 792.668454] env[61991]: _type = "Task" [ 792.668454] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.672958] env[61991]: INFO nova.scheduler.client.report [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Deleted allocations for instance 7df78da9-bf65-4621-b50d-43f1d721c2f1 [ 792.683969] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527e6ea2-386d-00d8-4eb8-14f1f5c9ddb1, 'name': SearchDatastore_Task, 'duration_secs': 0.010871} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.684853] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-938b1606-5eca-407b-bb0a-152e057e446c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.691217] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 792.691217] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eb7449-6f37-6640-a014-8524a75379e0" [ 792.691217] env[61991]: _type = "Task" [ 792.691217] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.701102] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eb7449-6f37-6640-a014-8524a75379e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.803133] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.851273] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.256226} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.852052] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.852628] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b57ce1-7ea5-4086-9b48-aac95714552e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.875386] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 3e62f219-7fec-470f-9d0a-e292d62e076c/3e62f219-7fec-470f-9d0a-e292d62e076c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.876716] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.877445] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.877784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.879370] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3694e7e-8436-4c94-b788-b61fafc4e48f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.895563] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2724fb0f-0596-4007-a15b-30061d01f14a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.918046] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.918511] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.918511] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.918663] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.918812] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.918958] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.919187] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.919350] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.919517] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.919681] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.919856] env[61991]: DEBUG nova.virt.hardware [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.927051] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfiguring VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 792.928030] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6861120f-cf01-46d9-a669-9e6e07134659 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.940714] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 792.940714] env[61991]: value = "task-1129692" [ 792.940714] env[61991]: _type = "Task" [ 792.940714] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.946453] env[61991]: DEBUG oslo_vmware.api [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 792.946453] env[61991]: value = "task-1129693" [ 792.946453] env[61991]: _type = "Task" [ 792.946453] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.953905] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129692, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.957155] env[61991]: DEBUG oslo_vmware.api [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129693, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.971081] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129691, 'name': ReconfigVM_Task, 'duration_secs': 0.310868} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.971081] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Reconfigured VM instance instance-00000040 to attach disk [datastore1] e3e5ec72-eec1-4155-bab2-038e3d5e38a8/e3e5ec72-eec1-4155-bab2-038e3d5e38a8.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.971742] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fbcf58b-42d9-43d4-998e-4ab38fcd7ee7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.978368] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 792.978368] env[61991]: value = "task-1129694" [ 792.978368] env[61991]: _type = "Task" [ 792.978368] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.989297] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129694, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.136486] env[61991]: DEBUG nova.compute.utils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.138130] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.138334] env[61991]: DEBUG nova.network.neutron [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 793.145418] env[61991]: DEBUG nova.compute.manager [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.146300] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bca187-b442-4d58-91fb-3f668c2bd5e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.181100] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e06a57e-3969-4920-a1ab-21ac2f3858c9 tempest-ServersNegativeTestMultiTenantJSON-718316372 tempest-ServersNegativeTestMultiTenantJSON-718316372-project-member] Lock "7df78da9-bf65-4621-b50d-43f1d721c2f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.973s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.200527] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eb7449-6f37-6640-a014-8524a75379e0, 'name': SearchDatastore_Task, 'duration_secs': 0.012027} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.200766] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.201037] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 884e465b-e14f-4114-81a6-40e2e8694558/884e465b-e14f-4114-81a6-40e2e8694558.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 793.201306] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d747521c-c8e6-403a-82a8-7c1ae8db272c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.208264] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 793.208264] env[61991]: value = "task-1129695" [ 793.208264] env[61991]: _type = "Task" [ 793.208264] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.215834] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129695, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.243188] env[61991]: DEBUG nova.policy [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b5d90a913fc41e282e050bb9af628d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adbd4ea9b0744f28a0d4a46b4a04d683', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.331115] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.454304] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129692, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.460473] env[61991]: DEBUG oslo_vmware.api [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129693, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.471486] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.487808] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129694, 'name': Rename_Task, 'duration_secs': 0.191991} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.488113] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 793.488376] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-031d136e-15d3-4fff-b807-926bb837b26d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.493993] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 793.493993] env[61991]: value = "task-1129696" [ 793.493993] env[61991]: _type = "Task" [ 793.493993] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.501902] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.637798] env[61991]: DEBUG nova.network.neutron [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Successfully updated port: c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.645321] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 793.656239] env[61991]: INFO nova.compute.manager [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] instance snapshotting [ 793.657565] env[61991]: DEBUG nova.objects.instance [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'flavor' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.721563] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129695, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.791390] env[61991]: DEBUG nova.network.neutron [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Successfully created port: fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.959132] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129692, 'name': ReconfigVM_Task, 'duration_secs': 0.889478} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.960129] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 3e62f219-7fec-470f-9d0a-e292d62e076c/3e62f219-7fec-470f-9d0a-e292d62e076c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.960615] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dfa7ad5f-8437-4c0e-9a3f-0edaf19a2675 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.965932] env[61991]: DEBUG oslo_vmware.api [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129693, 'name': ReconfigVM_Task, 'duration_secs': 0.553437} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.966728] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.966956] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfigured VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 793.974136] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 793.974136] env[61991]: value = "task-1129697" [ 793.974136] env[61991]: _type = "Task" [ 793.974136] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.982855] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129697, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.013765] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129696, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.045548] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01678f9-ac35-49c6-9459-2483080995f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.057372] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bb71db-d854-45fa-bacb-7a0bea762e06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.086157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfc92a9-95a3-4d88-89fd-8f8f0ecffd05 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.093996] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97314f91-c818-4ea4-8e10-05828a3c7988 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.107985] env[61991]: DEBUG nova.compute.provider_tree [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.140595] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.140927] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.141182] env[61991]: DEBUG nova.network.neutron [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 794.166741] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192b3b08-8ca9-4a99-9020-80df034907d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.188615] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea94fe45-a455-4aac-838b-84001d3003cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.218880] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129695, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688789} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.219233] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 884e465b-e14f-4114-81a6-40e2e8694558/884e465b-e14f-4114-81a6-40e2e8694558.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 794.219493] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.219791] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d9ee2b8-4f19-43df-831a-787589806360 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.228542] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 794.228542] env[61991]: value = "task-1129698" [ 794.228542] env[61991]: _type = "Task" [ 794.228542] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.237503] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129698, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.338581] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Received event network-vif-plugged-50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.338581] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Acquiring lock "884e465b-e14f-4114-81a6-40e2e8694558-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.338581] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Lock "884e465b-e14f-4114-81a6-40e2e8694558-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.338581] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Lock "884e465b-e14f-4114-81a6-40e2e8694558-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.338581] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] No waiting events found dispatching network-vif-plugged-50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 794.338581] env[61991]: WARNING nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Received unexpected event network-vif-plugged-50b6a76d-9313-4b7c-b577-09c086277647 for instance with vm_state building and task_state spawning. [ 794.338581] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Received event network-changed-50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.338581] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Refreshing instance network info cache due to event network-changed-50b6a76d-9313-4b7c-b577-09c086277647. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 794.338581] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Acquiring lock "refresh_cache-884e465b-e14f-4114-81a6-40e2e8694558" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.338581] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Acquired lock "refresh_cache-884e465b-e14f-4114-81a6-40e2e8694558" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.338581] env[61991]: DEBUG nova.network.neutron [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Refreshing network info cache for port 50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.477154] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5801e1e5-bdf4-4566-9a6d-7727c36f3455 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-e08c38be-49fa-4a29-9cb1-37e15930e258" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.007s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.486777] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129697, 'name': Rename_Task, 'duration_secs': 0.304465} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.487471] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.487848] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-058e3787-c57c-46a9-a94f-9671007f7398 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.496031] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 794.496031] env[61991]: value = "task-1129699" [ 794.496031] env[61991]: _type = "Task" [ 794.496031] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.506161] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129699, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.508646] env[61991]: DEBUG oslo_vmware.api [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129696, 'name': PowerOnVM_Task, 'duration_secs': 0.792105} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.509131] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.509467] env[61991]: INFO nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Took 14.06 seconds to spawn the instance on the hypervisor. [ 794.509921] env[61991]: DEBUG nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.511046] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da34369-c696-4aef-b08e-6fd513f99962 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.610828] env[61991]: DEBUG nova.scheduler.client.report [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.666371] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 794.691532] env[61991]: DEBUG nova.network.neutron [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 794.703201] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 794.706959] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e1c4ac4e-279e-4ebb-bf26-179dff9492aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.714373] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 794.714373] env[61991]: value = "task-1129700" [ 794.714373] env[61991]: _type = "Task" [ 794.714373] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.723538] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.724162] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.724541] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.724903] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.725181] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.725466] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.725946] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.726369] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.726704] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.727010] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.727336] env[61991]: DEBUG nova.virt.hardware [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.732183] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4488c8bd-a639-42e9-9db8-509e34caa748 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.747215] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129700, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.755441] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bcea09-ec0a-47c7-86a6-32f6f983a113 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.759886] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129698, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.25116} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.760391] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.761605] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2f4662-0b01-4270-8c6d-29a28dd051dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.795455] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 884e465b-e14f-4114-81a6-40e2e8694558/884e465b-e14f-4114-81a6-40e2e8694558.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.798581] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6efb25d-9281-4c52-8439-7124ab4ef902 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.820535] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 794.820535] env[61991]: value = "task-1129701" [ 794.820535] env[61991]: _type = "Task" [ 794.820535] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.830506] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129701, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.926448] env[61991]: DEBUG nova.network.neutron [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Updating instance_info_cache with network_info: [{"id": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "address": "fa:16:3e:ee:36:73", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9be477d-86", "ovs_interfaceid": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.008192] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129699, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.032765] env[61991]: INFO nova.compute.manager [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Took 48.51 seconds to build instance. [ 795.079559] env[61991]: DEBUG nova.network.neutron [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Updated VIF entry in instance network info cache for port 50b6a76d-9313-4b7c-b577-09c086277647. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 795.079998] env[61991]: DEBUG nova.network.neutron [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Updating instance_info_cache with network_info: [{"id": "50b6a76d-9313-4b7c-b577-09c086277647", "address": "fa:16:3e:35:98:fa", "network": {"id": "e3aa1719-f3cd-4ad6-bd6d-5f950bb7d8f2", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1572203683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5322f4e5522f4a3d8f915d056e11ebce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap50b6a76d-93", "ovs_interfaceid": "50b6a76d-9313-4b7c-b577-09c086277647", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.116610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.118811] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.120233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.829s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.123105] env[61991]: INFO nova.compute.claims [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.227040] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129700, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.331950] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129701, 'name': ReconfigVM_Task, 'duration_secs': 0.354905} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.335049] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 884e465b-e14f-4114-81a6-40e2e8694558/884e465b-e14f-4114-81a6-40e2e8694558.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.335049] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7a011b3-fa3a-4f27-84fe-b9024380d448 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.341022] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 795.341022] env[61991]: value = "task-1129702" [ 795.341022] env[61991]: _type = "Task" [ 795.341022] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.351815] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129702, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.432022] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.432022] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Instance network_info: |[{"id": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "address": "fa:16:3e:ee:36:73", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9be477d-86", "ovs_interfaceid": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 795.432022] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:36:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0df968ae-c1ef-4009-a0f4-6f2e799c2fda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9be477d-8647-4664-8ebf-1f1fe5e45492', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.441206] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Creating folder: Project (79c013b6c254403b91f520e2af18027f). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.441719] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d4589f0-586b-4d3e-9ac9-13abb19b1e6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.455029] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Created folder: Project (79c013b6c254403b91f520e2af18027f) in parent group-v246753. [ 795.455502] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Creating folder: Instances. Parent ref: group-v246867. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.455954] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83d8ce81-e023-47a2-b1f2-c431f19c9315 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.466417] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Created folder: Instances in parent group-v246867. [ 795.468052] env[61991]: DEBUG oslo.service.loopingcall [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.468052] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 795.468052] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a903c5d-665f-4f14-8c27-40bff5b8f391 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.492638] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.492638] env[61991]: value = "task-1129705" [ 795.492638] env[61991]: _type = "Task" [ 795.492638] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.506683] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129705, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.510373] env[61991]: DEBUG oslo_vmware.api [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129699, 'name': PowerOnVM_Task, 'duration_secs': 0.690462} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.510658] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.510912] env[61991]: INFO nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Took 9.39 seconds to spawn the instance on the hypervisor. [ 795.511186] env[61991]: DEBUG nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.512313] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974943f8-ebb7-4e48-b2ff-993d28eba8aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.537155] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe2faac9-fc2d-43c1-b076-ae8318f374b9 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.508s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.563983] env[61991]: DEBUG nova.network.neutron [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Successfully updated port: fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.588592] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Releasing lock "refresh_cache-884e465b-e14f-4114-81a6-40e2e8694558" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.588909] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-vif-plugged-e08c38be-49fa-4a29-9cb1-37e15930e258 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.589166] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.589726] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.589963] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.590201] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] No waiting events found dispatching network-vif-plugged-e08c38be-49fa-4a29-9cb1-37e15930e258 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 795.590418] env[61991]: WARNING nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received unexpected event network-vif-plugged-e08c38be-49fa-4a29-9cb1-37e15930e258 for instance with vm_state active and task_state None. [ 795.590624] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-changed-e08c38be-49fa-4a29-9cb1-37e15930e258 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.590866] env[61991]: DEBUG nova.compute.manager [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing instance network info cache due to event network-changed-e08c38be-49fa-4a29-9cb1-37e15930e258. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.591210] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.591488] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.591758] env[61991]: DEBUG nova.network.neutron [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Refreshing network info cache for port e08c38be-49fa-4a29-9cb1-37e15930e258 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.627800] env[61991]: DEBUG nova.compute.utils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.631684] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 795.631855] env[61991]: DEBUG nova.network.neutron [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 795.708514] env[61991]: DEBUG nova.policy [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8c995dbd884e95bd04e034e2999a56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e4cb3d1518d498a8cdc2aee3acb90cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 795.726558] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129700, 'name': CreateSnapshot_Task, 'duration_secs': 0.934712} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.726848] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 795.727652] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1edb53-b736-42b2-9fca-4022fa3b948d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.850916] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129702, 'name': Rename_Task, 'duration_secs': 0.17847} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.850916] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.850916] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1aa0d09-9cc2-4097-892c-499bdd20ca3f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.857340] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 795.857340] env[61991]: value = "task-1129706" [ 795.857340] env[61991]: _type = "Task" [ 795.857340] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.865705] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.002368] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129705, 'name': CreateVM_Task, 'duration_secs': 0.380016} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.002553] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 796.003284] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.003449] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.004381] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.004381] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5636cfc-1fb9-4762-8630-0a6f9fdc7f5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.010370] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 796.010370] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52107905-6389-9d8b-5210-9381221e338a" [ 796.010370] env[61991]: _type = "Task" [ 796.010370] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.018417] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52107905-6389-9d8b-5210-9381221e338a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.033540] env[61991]: INFO nova.compute.manager [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Took 44.78 seconds to build instance. [ 796.070264] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-49324fef-ad48-451b-a5ce-d9a1231137db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.070264] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-49324fef-ad48-451b-a5ce-d9a1231137db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.070264] env[61991]: DEBUG nova.network.neutron [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.137445] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.143758] env[61991]: DEBUG nova.network.neutron [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Successfully created port: f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.261148] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 796.272130] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ca5e0ab1-e58c-4c3d-a268-8eebdc1ec8ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.286859] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 796.286859] env[61991]: value = "task-1129707" [ 796.286859] env[61991]: _type = "Task" [ 796.286859] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.303474] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129707, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.337403] env[61991]: DEBUG nova.compute.manager [req-76a0cc28-ffc7-416c-8fa2-d4637a8e9c1d req-3456f422-d6a8-47a0-bcdc-ccbd0b9c3ab1 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Received event network-vif-plugged-fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.337637] env[61991]: DEBUG oslo_concurrency.lockutils [req-76a0cc28-ffc7-416c-8fa2-d4637a8e9c1d req-3456f422-d6a8-47a0-bcdc-ccbd0b9c3ab1 service nova] Acquiring lock "49324fef-ad48-451b-a5ce-d9a1231137db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.337887] env[61991]: DEBUG oslo_concurrency.lockutils [req-76a0cc28-ffc7-416c-8fa2-d4637a8e9c1d req-3456f422-d6a8-47a0-bcdc-ccbd0b9c3ab1 service nova] Lock "49324fef-ad48-451b-a5ce-d9a1231137db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.338012] env[61991]: DEBUG oslo_concurrency.lockutils [req-76a0cc28-ffc7-416c-8fa2-d4637a8e9c1d req-3456f422-d6a8-47a0-bcdc-ccbd0b9c3ab1 service nova] Lock "49324fef-ad48-451b-a5ce-d9a1231137db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.338450] env[61991]: DEBUG nova.compute.manager [req-76a0cc28-ffc7-416c-8fa2-d4637a8e9c1d req-3456f422-d6a8-47a0-bcdc-ccbd0b9c3ab1 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] No waiting events found dispatching network-vif-plugged-fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.338606] env[61991]: WARNING nova.compute.manager [req-76a0cc28-ffc7-416c-8fa2-d4637a8e9c1d req-3456f422-d6a8-47a0-bcdc-ccbd0b9c3ab1 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Received unexpected event network-vif-plugged-fa5c2434-a034-4ee0-b847-7108e636ef8d for instance with vm_state building and task_state spawning. [ 796.373654] env[61991]: DEBUG oslo_vmware.api [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129706, 'name': PowerOnVM_Task, 'duration_secs': 0.460516} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.373933] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.374982] env[61991]: INFO nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Took 7.65 seconds to spawn the instance on the hypervisor. [ 796.374982] env[61991]: DEBUG nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.376027] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997d75eb-2763-432f-af98-a1d81d964db5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.407598] env[61991]: DEBUG nova.network.neutron [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updated VIF entry in instance network info cache for port e08c38be-49fa-4a29-9cb1-37e15930e258. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.408271] env[61991]: DEBUG nova.network.neutron [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "address": "fa:16:3e:5c:e6:a5", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa20e94c9-8e", "ovs_interfaceid": "a20e94c9-8ee3-4696-b020-423d80f00dc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e08c38be-49fa-4a29-9cb1-37e15930e258", "address": "fa:16:3e:41:e8:b9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape08c38be-49", "ovs_interfaceid": "e08c38be-49fa-4a29-9cb1-37e15930e258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.418622] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.418622] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.419010] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.419223] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.419400] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.423061] env[61991]: INFO nova.compute.manager [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Terminating instance [ 796.425668] env[61991]: DEBUG nova.compute.manager [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.426382] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.428596] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d534fda0-8539-4da8-a3fd-9f44689a5e5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.444776] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.445203] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ff504e1-103d-4180-bea7-140c83434623 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.455028] env[61991]: DEBUG oslo_vmware.api [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 796.455028] env[61991]: value = "task-1129708" [ 796.455028] env[61991]: _type = "Task" [ 796.455028] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.469670] env[61991]: DEBUG oslo_vmware.api [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.524469] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52107905-6389-9d8b-5210-9381221e338a, 'name': SearchDatastore_Task, 'duration_secs': 0.014055} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.524867] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.525222] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.525353] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.525495] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.525758] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.525952] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42b0217a-ad10-49eb-b2c1-940cf2d2869c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.536362] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3691907-df45-45ad-83dc-97b1f740f46c tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.075s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.538176] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.538395] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.539206] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4ac1952-85b6-4e12-8481-f5db68b2a20c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.545175] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 796.545175] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5266f495-7247-6746-b728-68b20c2bfc5f" [ 796.545175] env[61991]: _type = "Task" [ 796.545175] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.558279] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5266f495-7247-6746-b728-68b20c2bfc5f, 'name': SearchDatastore_Task, 'duration_secs': 0.010045} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.561352] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a424c72-7a37-4d2b-9bcd-908ae02414c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.567362] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 796.567362] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52af8131-44d1-2b5a-720c-1efcbae34b33" [ 796.567362] env[61991]: _type = "Task" [ 796.567362] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.577921] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52af8131-44d1-2b5a-720c-1efcbae34b33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.594808] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf32944-24e2-41f2-9f9a-149d7b4bbaa6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.602966] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813f1bff-627c-46a0-b3a7-13556ce9612d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.646664] env[61991]: DEBUG nova.network.neutron [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.649974] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cec5b08-f3e1-4e94-bb0d-48fe50fd5439 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.659031] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f89a37e-e615-45a4-9c97-61b9223d4a32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.678538] env[61991]: DEBUG nova.compute.provider_tree [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.800645] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129707, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.826470] env[61991]: DEBUG nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Received event network-vif-plugged-c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.826660] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Acquiring lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.826887] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.827075] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.827245] env[61991]: DEBUG nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] No waiting events found dispatching network-vif-plugged-c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.827409] env[61991]: WARNING nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Received unexpected event network-vif-plugged-c9be477d-8647-4664-8ebf-1f1fe5e45492 for instance with vm_state building and task_state spawning. [ 796.827668] env[61991]: DEBUG nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Received event network-changed-c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.827849] env[61991]: DEBUG nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Refreshing instance network info cache due to event network-changed-c9be477d-8647-4664-8ebf-1f1fe5e45492. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 796.828050] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Acquiring lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.828196] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Acquired lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.828338] env[61991]: DEBUG nova.network.neutron [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Refreshing network info cache for port c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 796.896190] env[61991]: INFO nova.compute.manager [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Took 40.95 seconds to build instance. [ 796.918912] env[61991]: DEBUG oslo_concurrency.lockutils [req-0d65e07c-93ea-42dc-8811-d25ae74227ae req-e5d8c218-500b-44b4-856f-b70667d5c49f service nova] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.926033] env[61991]: DEBUG nova.network.neutron [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Updating instance_info_cache with network_info: [{"id": "fa5c2434-a034-4ee0-b847-7108e636ef8d", "address": "fa:16:3e:1f:c3:39", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa5c2434-a0", "ovs_interfaceid": "fa5c2434-a034-4ee0-b847-7108e636ef8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.938113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "c14d99a3-950d-44d0-b330-3031a1f8a2be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.938113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.938113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.938113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.938113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.939722] env[61991]: INFO nova.compute.manager [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Terminating instance [ 796.947634] env[61991]: DEBUG nova.compute.manager [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.947634] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.948448] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3f3d0d-e2a3-49e9-b025-f0c5ea6ea128 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.957359] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.960918] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f05d41fb-9b9c-4ab1-a1f3-1f975f57e337 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.968238] env[61991]: DEBUG oslo_vmware.api [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129708, 'name': PowerOffVM_Task, 'duration_secs': 0.437207} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.970802] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.970802] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 796.970935] env[61991]: DEBUG oslo_vmware.api [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 796.970935] env[61991]: value = "task-1129709" [ 796.970935] env[61991]: _type = "Task" [ 796.970935] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.971497] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7a3814d-2085-403f-a6cd-a7bebb62a852 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.987740] env[61991]: DEBUG oslo_vmware.api [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.062813] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.063099] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.063301] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleting the datastore file [datastore1] e3e5ec72-eec1-4155-bab2-038e3d5e38a8 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.063657] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ac80a03-d3b5-4c06-95db-0b775bba4a70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.074582] env[61991]: DEBUG oslo_vmware.api [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 797.074582] env[61991]: value = "task-1129711" [ 797.074582] env[61991]: _type = "Task" [ 797.074582] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.082599] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52af8131-44d1-2b5a-720c-1efcbae34b33, 'name': SearchDatastore_Task, 'duration_secs': 0.009671} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.083318] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.083913] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/f6945e87-3b17-4e0e-845f-3bacf0fbe5ac.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 797.083913] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e44d7840-5db1-47a6-aee1-082d2a6d36ef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.094132] env[61991]: DEBUG oslo_vmware.api [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.099365] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 797.099365] env[61991]: value = "task-1129712" [ 797.099365] env[61991]: _type = "Task" [ 797.099365] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.108238] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.151022] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.181405] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.181686] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.181874] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.183443] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.183806] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.184085] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.184730] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.185072] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.185488] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.186271] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.186499] env[61991]: DEBUG nova.virt.hardware [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.188145] env[61991]: DEBUG nova.scheduler.client.report [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.193031] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14b81e8-a1cd-4029-a619-66644e6514f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.202506] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b4893a-f8ad-4f55-bb98-969c3cf18697 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.300490] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129707, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.399335] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4b7cb600-b69d-4e47-b05b-0e8ea110d670 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "884e465b-e14f-4114-81a6-40e2e8694558" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.829s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.429150] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-49324fef-ad48-451b-a5ce-d9a1231137db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.429150] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance network_info: |[{"id": "fa5c2434-a034-4ee0-b847-7108e636ef8d", "address": "fa:16:3e:1f:c3:39", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa5c2434-a0", "ovs_interfaceid": "fa5c2434-a034-4ee0-b847-7108e636ef8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 797.429288] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:c3:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa5c2434-a034-4ee0-b847-7108e636ef8d', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.437740] env[61991]: DEBUG oslo.service.loopingcall [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.440868] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.441334] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17a1aa6d-cb10-41bb-9571-da4b26c2ba61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.466880] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.466880] env[61991]: value = "task-1129713" [ 797.466880] env[61991]: _type = "Task" [ 797.466880] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.481370] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129713, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.489400] env[61991]: DEBUG oslo_vmware.api [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129709, 'name': PowerOffVM_Task, 'duration_secs': 0.285397} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.489762] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.489977] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.490277] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d07ea816-9d42-4e41-b507-d33427f512fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.567730] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.567730] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.567730] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Deleting the datastore file [datastore1] c14d99a3-950d-44d0-b330-3031a1f8a2be {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.567730] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94ab7483-9f14-4ba1-97d0-ccb29dcb5fe7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.577148] env[61991]: DEBUG oslo_vmware.api [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for the task: (returnval){ [ 797.577148] env[61991]: value = "task-1129715" [ 797.577148] env[61991]: _type = "Task" [ 797.577148] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.592797] env[61991]: DEBUG oslo_vmware.api [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190522} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.597963] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.600636] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.601170] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.601579] env[61991]: INFO nova.compute.manager [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Took 1.18 seconds to destroy the instance on the hypervisor. [ 797.602042] env[61991]: DEBUG oslo.service.loopingcall [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.604075] env[61991]: DEBUG oslo_vmware.api [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129715, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.604075] env[61991]: DEBUG nova.compute.manager [-] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.604075] env[61991]: DEBUG nova.network.neutron [-] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.614957] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129712, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.697702] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.699199] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.703443] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.430s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.704980] env[61991]: DEBUG nova.objects.instance [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lazy-loading 'resources' on Instance uuid d2a0e5d1-5557-41b6-936d-dc86d1346c61 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 797.732566] env[61991]: DEBUG nova.network.neutron [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Updated VIF entry in instance network info cache for port c9be477d-8647-4664-8ebf-1f1fe5e45492. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 797.733069] env[61991]: DEBUG nova.network.neutron [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Updating instance_info_cache with network_info: [{"id": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "address": "fa:16:3e:ee:36:73", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9be477d-86", "ovs_interfaceid": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.802947] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129707, 'name': CloneVM_Task, 'duration_secs': 1.184134} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.803262] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created linked-clone VM from snapshot [ 797.804357] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde726d4-ac81-4cd8-8b03-86947443a54d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.819116] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Uploading image add4fbc4-635d-4eaf-945a-ac91215c88f3 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 797.849821] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 797.849821] env[61991]: value = "vm-246871" [ 797.849821] env[61991]: _type = "VirtualMachine" [ 797.849821] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 797.850503] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5a2fc468-2785-44fb-a151-a20f3fce9621 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.858453] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease: (returnval){ [ 797.858453] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521dc7ab-cc3b-0eaf-af81-45547bc83f5c" [ 797.858453] env[61991]: _type = "HttpNfcLease" [ 797.858453] env[61991]: } obtained for exporting VM: (result){ [ 797.858453] env[61991]: value = "vm-246871" [ 797.858453] env[61991]: _type = "VirtualMachine" [ 797.858453] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 797.858773] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the lease: (returnval){ [ 797.858773] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521dc7ab-cc3b-0eaf-af81-45547bc83f5c" [ 797.858773] env[61991]: _type = "HttpNfcLease" [ 797.858773] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 797.866030] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 797.866030] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521dc7ab-cc3b-0eaf-af81-45547bc83f5c" [ 797.866030] env[61991]: _type = "HttpNfcLease" [ 797.866030] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 797.976949] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129713, 'name': CreateVM_Task, 'duration_secs': 0.462392} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.978601] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.978601] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.978601] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.978601] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.978601] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-298d74e0-0aed-4e98-af05-2392c5ef7ae8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.983631] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 797.983631] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b6fec-b060-9b04-0010-efc5e8ddaeee" [ 797.983631] env[61991]: _type = "Task" [ 797.983631] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.991709] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b6fec-b060-9b04-0010-efc5e8ddaeee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.094244] env[61991]: DEBUG oslo_vmware.api [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Task: {'id': task-1129715, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24913} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.094552] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.094806] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 798.095088] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 798.095287] env[61991]: INFO nova.compute.manager [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Took 1.15 seconds to destroy the instance on the hypervisor. [ 798.095560] env[61991]: DEBUG oslo.service.loopingcall [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.095786] env[61991]: DEBUG nova.compute.manager [-] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.095884] env[61991]: DEBUG nova.network.neutron [-] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.119736] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643458} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.120052] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/f6945e87-3b17-4e0e-845f-3bacf0fbe5ac.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 798.120283] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.120545] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9dcd6224-7f4d-4603-b219-8dd1a7df0116 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.127973] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 798.127973] env[61991]: value = "task-1129717" [ 798.127973] env[61991]: _type = "Task" [ 798.127973] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.140028] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129717, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.209847] env[61991]: DEBUG nova.compute.utils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.215279] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.215791] env[61991]: DEBUG nova.network.neutron [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 798.235901] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Releasing lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.236268] env[61991]: DEBUG nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Received event network-changed-fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.236498] env[61991]: DEBUG nova.compute.manager [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Refreshing instance network info cache due to event network-changed-fa5c2434-a034-4ee0-b847-7108e636ef8d. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 798.236793] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Acquiring lock "refresh_cache-49324fef-ad48-451b-a5ce-d9a1231137db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.236989] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Acquired lock "refresh_cache-49324fef-ad48-451b-a5ce-d9a1231137db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.237200] env[61991]: DEBUG nova.network.neutron [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Refreshing network info cache for port fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.298485] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-a20e94c9-8ee3-4696-b020-423d80f00dc9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.298609] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-a20e94c9-8ee3-4696-b020-423d80f00dc9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.346058] env[61991]: DEBUG nova.policy [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceec39441db345a1ac9cc3f9d6f6ee18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '870d2c2c0e554180b190b88bdab5fc2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.371225] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 798.371225] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521dc7ab-cc3b-0eaf-af81-45547bc83f5c" [ 798.371225] env[61991]: _type = "HttpNfcLease" [ 798.371225] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 798.371225] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 798.371225] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521dc7ab-cc3b-0eaf-af81-45547bc83f5c" [ 798.371225] env[61991]: _type = "HttpNfcLease" [ 798.371225] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 798.371925] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312dfd84-ee85-4edf-9be5-5ad3d2d9abf1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.379874] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5218d801-4ab0-b8cf-8b00-ba41e7d92309/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 798.380539] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5218d801-4ab0-b8cf-8b00-ba41e7d92309/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 798.454883] env[61991]: DEBUG nova.network.neutron [-] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.497338] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b6fec-b060-9b04-0010-efc5e8ddaeee, 'name': SearchDatastore_Task, 'duration_secs': 0.009616} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.503316] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.503575] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.503808] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.504091] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.504284] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.506164] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dda145e4-2bbb-43ff-a632-34e5f832bc8b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.507166] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-88ce440e-d920-4e6b-b2f8-d97f39ffae96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.519477] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.519672] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.520605] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c588608-15c8-4099-b45e-cb06129f0130 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.526373] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 798.526373] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520845a4-076f-6d3d-9f60-ec4728e60517" [ 798.526373] env[61991]: _type = "Task" [ 798.526373] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.538174] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520845a4-076f-6d3d-9f60-ec4728e60517, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.606623] env[61991]: DEBUG nova.network.neutron [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Successfully updated port: f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.640371] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129717, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.263253} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.643404] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.645128] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b767f5-0812-4837-95ff-b22a4c082549 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.664563] env[61991]: DEBUG nova.network.neutron [-] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.676572] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/f6945e87-3b17-4e0e-845f-3bacf0fbe5ac.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.679854] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e9de7f8-7e1f-4db0-bf7f-1b2086f0a386 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.695926] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b7ea2e-3adb-4f49-bead-722853862338 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.706109] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d82193a-2a7d-4d62-b6e5-28c0642daddb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.709179] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 798.709179] env[61991]: value = "task-1129718" [ 798.709179] env[61991]: _type = "Task" [ 798.709179] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.739040] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.748229] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64389f33-51f5-4cb3-9c4b-45bf1ac4b93d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.759393] env[61991]: DEBUG nova.compute.manager [req-c7fceadf-f154-4dac-8c3d-0cf67a917d82 req-13b3d3b5-1aa0-42ad-9a4c-bab5527cde85 service nova] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Received event network-vif-deleted-863ee2c0-0c5a-49c8-a443-3c1c3d73c458 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.760229] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129718, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.768623] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da42a626-15e1-474b-846b-58ca8e8bee1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.782645] env[61991]: DEBUG nova.compute.provider_tree [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.805399] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.805399] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.805481] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84971e90-dabe-4948-a176-201478fb38ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.824826] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf593157-395f-48a3-9d93-464b2c23a0f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.857698] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfiguring VM to detach interface {{(pid=61991) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 798.860418] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01e6408d-6dbc-449b-8188-5d16e088d350 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.882315] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 798.882315] env[61991]: value = "task-1129719" [ 798.882315] env[61991]: _type = "Task" [ 798.882315] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.889390] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.958400] env[61991]: INFO nova.compute.manager [-] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Took 0.86 seconds to deallocate network for instance. [ 799.025769] env[61991]: DEBUG nova.compute.manager [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Received event network-vif-deleted-2d37704c-b54b-459f-b593-12a9062c3601 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.029315] env[61991]: DEBUG nova.compute.manager [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Received event network-vif-plugged-f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.029315] env[61991]: DEBUG oslo_concurrency.lockutils [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] Acquiring lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.029315] env[61991]: DEBUG oslo_concurrency.lockutils [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.029315] env[61991]: DEBUG oslo_concurrency.lockutils [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.029507] env[61991]: DEBUG nova.compute.manager [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] No waiting events found dispatching network-vif-plugged-f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.030307] env[61991]: WARNING nova.compute.manager [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Received unexpected event network-vif-plugged-f9418974-50f5-48bc-a8cc-dd016b21728a for instance with vm_state building and task_state spawning. [ 799.030482] env[61991]: DEBUG nova.compute.manager [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Received event network-changed-f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.030679] env[61991]: DEBUG nova.compute.manager [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Refreshing instance network info cache due to event network-changed-f9418974-50f5-48bc-a8cc-dd016b21728a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.030889] env[61991]: DEBUG oslo_concurrency.lockutils [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] Acquiring lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.031028] env[61991]: DEBUG oslo_concurrency.lockutils [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] Acquired lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.031233] env[61991]: DEBUG nova.network.neutron [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Refreshing network info cache for port f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 799.068517] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520845a4-076f-6d3d-9f60-ec4728e60517, 'name': SearchDatastore_Task, 'duration_secs': 0.011178} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.069556] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0614041c-58b6-413f-a636-a24c571d802b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.077985] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 799.077985] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5200e91f-6e47-06f7-bd6d-32636a2a003e" [ 799.077985] env[61991]: _type = "Task" [ 799.077985] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.094362] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5200e91f-6e47-06f7-bd6d-32636a2a003e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.109203] env[61991]: DEBUG nova.network.neutron [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Successfully created port: 9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.115143] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.166758] env[61991]: INFO nova.compute.manager [-] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Took 1.56 seconds to deallocate network for instance. [ 799.221646] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129718, 'name': ReconfigVM_Task, 'duration_secs': 0.410256} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.222124] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Reconfigured VM instance instance-00000043 to attach disk [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/f6945e87-3b17-4e0e-845f-3bacf0fbe5ac.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.223059] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3aab8a5e-9bf4-4826-a9f3-45b377f102ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.230419] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 799.230419] env[61991]: value = "task-1129720" [ 799.230419] env[61991]: _type = "Task" [ 799.230419] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.239910] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129720, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.286502] env[61991]: DEBUG nova.scheduler.client.report [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.353689] env[61991]: DEBUG nova.compute.manager [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.354708] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4668df-02e5-496e-830c-6c6cd6688dd8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.390930] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.467391] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.546130] env[61991]: DEBUG nova.network.neutron [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Updated VIF entry in instance network info cache for port fa5c2434-a034-4ee0-b847-7108e636ef8d. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.546572] env[61991]: DEBUG nova.network.neutron [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Updating instance_info_cache with network_info: [{"id": "fa5c2434-a034-4ee0-b847-7108e636ef8d", "address": "fa:16:3e:1f:c3:39", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa5c2434-a0", "ovs_interfaceid": "fa5c2434-a034-4ee0-b847-7108e636ef8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.601713] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5200e91f-6e47-06f7-bd6d-32636a2a003e, 'name': SearchDatastore_Task, 'duration_secs': 0.011688} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.602815] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.603175] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.603800] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-052345e4-dd59-4e4c-a4df-75032f9251f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.616432] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 799.616432] env[61991]: value = "task-1129721" [ 799.616432] env[61991]: _type = "Task" [ 799.616432] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.630854] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129721, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.675237] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.683930] env[61991]: DEBUG nova.network.neutron [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.743742] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129720, 'name': Rename_Task, 'duration_secs': 0.173638} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.744352] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.745134] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dfbf8f9-f078-416c-89c2-bb5a8f2ad843 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.750441] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.755358] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 799.755358] env[61991]: value = "task-1129722" [ 799.755358] env[61991]: _type = "Task" [ 799.755358] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.767520] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129722, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.785151] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.785456] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.785731] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.785909] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.786175] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.786498] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.786737] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.786931] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.787179] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.787438] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.787644] env[61991]: DEBUG nova.virt.hardware [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.788533] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a58a64-7e87-47f8-8b53-f2fba5c3bd35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.791746] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.089s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.794827] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.949s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.800075] env[61991]: DEBUG nova.objects.instance [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lazy-loading 'resources' on Instance uuid d6a2d226-5097-4461-b2ce-fa698f1066d5 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.801945] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4238e6-b69c-4dd9-8fad-f199cb0e7754 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.822824] env[61991]: INFO nova.scheduler.client.report [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleted allocations for instance d2a0e5d1-5557-41b6-936d-dc86d1346c61 [ 799.867432] env[61991]: INFO nova.compute.manager [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] instance snapshotting [ 799.870975] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfe7cbd-a0c7-42cb-8138-b7acf47999df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.900788] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6473e84-2c15-4d6d-99a9-e4d10a84974e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.904534] env[61991]: DEBUG nova.network.neutron [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.912568] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.049538] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce2f1716-9281-4482-9f96-5b73f396f143 req-6b67e069-196e-4fa1-bb46-34be399e78b6 service nova] Releasing lock "refresh_cache-49324fef-ad48-451b-a5ce-d9a1231137db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.131739] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129721, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.267686] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129722, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.336513] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5ee4bcd2-3142-47cf-bfd6-9480549038ed tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "d2a0e5d1-5557-41b6-936d-dc86d1346c61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.193s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.409655] env[61991]: DEBUG oslo_concurrency.lockutils [req-f4f59c6b-bf20-4c9e-9d03-566a54e4af6c req-93a3088c-8c9a-4c28-bc70-d5dcb4ad07f7 service nova] Releasing lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.409973] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.410970] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.410970] env[61991]: DEBUG nova.network.neutron [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.419949] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 800.420283] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b5a31333-308d-4e4a-9e08-6f243a4494f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.431136] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 800.431136] env[61991]: value = "task-1129723" [ 800.431136] env[61991]: _type = "Task" [ 800.431136] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.440155] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129723, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.629037] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129721, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625462} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.629319] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 800.629554] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 800.629810] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79e878b1-c044-447c-804c-d2d26cea9882 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.638211] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 800.638211] env[61991]: value = "task-1129724" [ 800.638211] env[61991]: _type = "Task" [ 800.638211] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.649147] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.659570] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb6e34f-93d4-4b68-a459-f6e17c86da51 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.669127] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26643c2-4f7b-41e0-af7d-31cf3078a823 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.707659] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae5af0e-82d3-4b3e-8ef6-fbf2f062a9f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.715746] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b36223f-3a95-4e65-a20e-44f9e845d7de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.734896] env[61991]: DEBUG nova.compute.provider_tree [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.767372] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129722, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.909938] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.946147] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129723, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.060942] env[61991]: DEBUG nova.network.neutron [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.152029] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076263} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.156970] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.158481] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e691c91-0b2a-4f07-8a9b-588333fb444f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.182178] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.186141] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3736a265-f58a-4d51-b73c-1807e6206548 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.208849] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 801.208849] env[61991]: value = "task-1129725" [ 801.208849] env[61991]: _type = "Task" [ 801.208849] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.217706] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129725, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.241320] env[61991]: DEBUG nova.scheduler.client.report [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.243716] env[61991]: DEBUG nova.network.neutron [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Updating instance_info_cache with network_info: [{"id": "f9418974-50f5-48bc-a8cc-dd016b21728a", "address": "fa:16:3e:f8:d8:1a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9418974-50", "ovs_interfaceid": "f9418974-50f5-48bc-a8cc-dd016b21728a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.273198] env[61991]: DEBUG oslo_vmware.api [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129722, 'name': PowerOnVM_Task, 'duration_secs': 1.146214} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.273198] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.273198] env[61991]: INFO nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Took 9.08 seconds to spawn the instance on the hypervisor. [ 801.273198] env[61991]: DEBUG nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.274365] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84430bd5-c66b-4109-88dd-a8f77a937bd9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.404863] env[61991]: DEBUG nova.compute.manager [req-e57bad0c-8c73-4ac2-9d24-5f82c80da31e req-071006cd-a544-4cd9-82be-2fddcccfe509 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Received event network-vif-plugged-9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.405308] env[61991]: DEBUG oslo_concurrency.lockutils [req-e57bad0c-8c73-4ac2-9d24-5f82c80da31e req-071006cd-a544-4cd9-82be-2fddcccfe509 service nova] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.405688] env[61991]: DEBUG oslo_concurrency.lockutils [req-e57bad0c-8c73-4ac2-9d24-5f82c80da31e req-071006cd-a544-4cd9-82be-2fddcccfe509 service nova] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.406069] env[61991]: DEBUG oslo_concurrency.lockutils [req-e57bad0c-8c73-4ac2-9d24-5f82c80da31e req-071006cd-a544-4cd9-82be-2fddcccfe509 service nova] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.406751] env[61991]: DEBUG nova.compute.manager [req-e57bad0c-8c73-4ac2-9d24-5f82c80da31e req-071006cd-a544-4cd9-82be-2fddcccfe509 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] No waiting events found dispatching network-vif-plugged-9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 801.409058] env[61991]: WARNING nova.compute.manager [req-e57bad0c-8c73-4ac2-9d24-5f82c80da31e req-071006cd-a544-4cd9-82be-2fddcccfe509 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Received unexpected event network-vif-plugged-9452191d-e715-4ad9-a48e-43aebb90b551 for instance with vm_state building and task_state spawning. [ 801.411425] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.449339] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129723, 'name': CreateSnapshot_Task, 'duration_secs': 0.871646} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.449936] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 801.450871] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3f979f-4f4b-45ba-882c-6b04de94b620 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.496867] env[61991]: DEBUG nova.network.neutron [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Successfully updated port: 9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.719337] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129725, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.744879] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.950s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.748367] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.497s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.748367] env[61991]: DEBUG nova.objects.instance [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lazy-loading 'resources' on Instance uuid 0aa67bb1-59d3-42e4-8f12-47e61a130dfc {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.752155] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.752155] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Instance network_info: |[{"id": "f9418974-50f5-48bc-a8cc-dd016b21728a", "address": "fa:16:3e:f8:d8:1a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9418974-50", "ovs_interfaceid": "f9418974-50f5-48bc-a8cc-dd016b21728a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.752155] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:d8:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9418974-50f5-48bc-a8cc-dd016b21728a', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.764131] env[61991]: DEBUG oslo.service.loopingcall [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.764727] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.765273] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d09260ee-1a84-4ef8-84b2-1bffe5a91d50 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.786858] env[61991]: INFO nova.scheduler.client.report [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Deleted allocations for instance d6a2d226-5097-4461-b2ce-fa698f1066d5 [ 801.804393] env[61991]: INFO nova.compute.manager [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Took 39.90 seconds to build instance. [ 801.808068] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.808068] env[61991]: value = "task-1129726" [ 801.808068] env[61991]: _type = "Task" [ 801.808068] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.819882] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129726, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.909971] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.974885] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 801.974885] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-459e88d1-cfe6-4388-ad9f-5412777932ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.984410] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 801.984410] env[61991]: value = "task-1129727" [ 801.984410] env[61991]: _type = "Task" [ 801.984410] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.993227] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129727, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.998063] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.998063] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.998200] env[61991]: DEBUG nova.network.neutron [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.222089] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129725, 'name': ReconfigVM_Task, 'duration_secs': 0.618684} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.222620] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.223601] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12bea419-96f2-4fdd-be1a-845111471784 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.232385] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 802.232385] env[61991]: value = "task-1129728" [ 802.232385] env[61991]: _type = "Task" [ 802.232385] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.248207] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129728, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.298496] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c327c8c2-8fb1-4c53-abfe-0d6be237ca86 tempest-FloatingIPsAssociationNegativeTestJSON-139780541 tempest-FloatingIPsAssociationNegativeTestJSON-139780541-project-member] Lock "d6a2d226-5097-4461-b2ce-fa698f1066d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.444s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.307553] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5830e2-06b2-4278-a149-7de9d43f4a7b tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.807s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.322403] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129726, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.413809] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.498462] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129727, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.547430] env[61991]: DEBUG nova.network.neutron [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.631968] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1b9243-6391-4f76-9f05-d0e771ebe7f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.642416] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a933103-f4d8-4ef7-95bf-ba89e57dd3a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.677098] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2e0556-ed0f-4b3a-9262-d65df7544647 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.685412] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5cd0dd-dd21-490c-bb5a-5f59713b7a60 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.703412] env[61991]: DEBUG nova.compute.provider_tree [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.745039] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129728, 'name': Rename_Task, 'duration_secs': 0.24949} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.745745] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.746248] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c45c9aab-c8a7-495a-a6b6-8520576f2773 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.752844] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 802.752844] env[61991]: value = "task-1129729" [ 802.752844] env[61991]: _type = "Task" [ 802.752844] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.762329] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.794114] env[61991]: DEBUG nova.network.neutron [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.818819] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129726, 'name': CreateVM_Task, 'duration_secs': 0.690298} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.818987] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.819768] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.819880] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.820233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.820744] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-346e1741-2330-463b-8ee6-ecdde202605c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.826308] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 802.826308] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205e0e5-c468-14fc-c046-13d557b08e45" [ 802.826308] env[61991]: _type = "Task" [ 802.826308] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.835404] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205e0e5-c468-14fc-c046-13d557b08e45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.909428] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.995289] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129727, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.207064] env[61991]: DEBUG nova.scheduler.client.report [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.270469] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129729, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.298970] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.299149] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Instance network_info: |[{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 803.299657] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:8a:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9452191d-e715-4ad9-a48e-43aebb90b551', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.309472] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating folder: Project (870d2c2c0e554180b190b88bdab5fc2d). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.310721] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c34bed17-666b-4c2d-9fe4-4bfbfa6dbc84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.324069] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created folder: Project (870d2c2c0e554180b190b88bdab5fc2d) in parent group-v246753. [ 803.324443] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating folder: Instances. Parent ref: group-v246876. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.324901] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0224f92-ad74-447f-a421-11bdad4875d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.337757] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205e0e5-c468-14fc-c046-13d557b08e45, 'name': SearchDatastore_Task, 'duration_secs': 0.022431} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.338482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.338482] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.338663] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.338826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.338997] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.339349] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f002b69-95e1-4361-bfff-95823c3c6812 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.343570] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created folder: Instances in parent group-v246876. [ 803.343848] env[61991]: DEBUG oslo.service.loopingcall [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.344536] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.344802] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-881480be-df3d-4f54-8d69-db4fd72adac0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.363903] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.364168] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 803.365542] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f99ceb9-7e07-4ebf-a420-f749a875b9d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.373055] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 803.373055] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c7b24d-e9d2-c435-0910-01778c4ac79b" [ 803.373055] env[61991]: _type = "Task" [ 803.373055] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.373501] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.373501] env[61991]: value = "task-1129732" [ 803.373501] env[61991]: _type = "Task" [ 803.373501] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.389317] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c7b24d-e9d2-c435-0910-01778c4ac79b, 'name': SearchDatastore_Task, 'duration_secs': 0.012971} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.393852] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129732, 'name': CreateVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.394209] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c613cae2-3d7b-4fdb-b95a-7c4979e506eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.405879] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 803.405879] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52753a62-ce3d-4266-5b40-d39a837e3277" [ 803.405879] env[61991]: _type = "Task" [ 803.405879] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.416051] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.422297] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52753a62-ce3d-4266-5b40-d39a837e3277, 'name': SearchDatastore_Task, 'duration_secs': 0.015952} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.422699] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.423019] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 62b3e5ae-b90c-47f8-95c4-14587dbf647d/62b3e5ae-b90c-47f8-95c4-14587dbf647d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.423325] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d13cd1d7-0d0a-4a3a-8269-5189221ff125 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.430813] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 803.430813] env[61991]: value = "task-1129733" [ 803.430813] env[61991]: _type = "Task" [ 803.430813] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.440363] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.497184] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129727, 'name': CloneVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.588028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "884e465b-e14f-4114-81a6-40e2e8694558" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.588028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "884e465b-e14f-4114-81a6-40e2e8694558" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.588028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "884e465b-e14f-4114-81a6-40e2e8694558-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.588028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "884e465b-e14f-4114-81a6-40e2e8694558-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.588028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "884e465b-e14f-4114-81a6-40e2e8694558-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.595629] env[61991]: INFO nova.compute.manager [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Terminating instance [ 803.600026] env[61991]: DEBUG nova.compute.manager [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 803.600026] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.600026] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7fa9de-d138-4e8e-ad05-d82f476b1167 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.614603] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 803.615225] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f5f82d1-630d-4e74-bb9b-e33165eb9a7f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.625205] env[61991]: DEBUG oslo_vmware.api [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 803.625205] env[61991]: value = "task-1129734" [ 803.625205] env[61991]: _type = "Task" [ 803.625205] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.636641] env[61991]: DEBUG oslo_vmware.api [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.712041] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.964s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.715283] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.409s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.716820] env[61991]: DEBUG nova.objects.instance [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lazy-loading 'resources' on Instance uuid f78ef63b-453e-45d3-959b-4b0c1922b53e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 803.762438] env[61991]: INFO nova.scheduler.client.report [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted allocations for instance 0aa67bb1-59d3-42e4-8f12-47e61a130dfc [ 803.769618] env[61991]: INFO nova.compute.manager [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Rescuing [ 803.770505] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.770826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.771150] env[61991]: DEBUG nova.network.neutron [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.784328] env[61991]: DEBUG oslo_vmware.api [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129729, 'name': PowerOnVM_Task, 'duration_secs': 0.874141} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.785311] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.785311] env[61991]: INFO nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Took 9.12 seconds to spawn the instance on the hypervisor. [ 803.785479] env[61991]: DEBUG nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.787354] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4c8785-024a-4975-82da-9edd21d75738 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.895204] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129732, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.917047] env[61991]: DEBUG nova.compute.manager [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Received event network-changed-9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.917047] env[61991]: DEBUG nova.compute.manager [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Refreshing instance network info cache due to event network-changed-9452191d-e715-4ad9-a48e-43aebb90b551. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 803.917288] env[61991]: DEBUG oslo_concurrency.lockutils [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.917661] env[61991]: DEBUG oslo_concurrency.lockutils [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.917661] env[61991]: DEBUG nova.network.neutron [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Refreshing network info cache for port 9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 803.928888] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.944143] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129733, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.001477] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129727, 'name': CloneVM_Task, 'duration_secs': 1.524063} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.001746] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Created linked-clone VM from snapshot [ 804.002600] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184f70b7-fe85-41fb-8624-cbbba9d7c80c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.013867] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Uploading image e5b4302d-ea69-40e0-9557-d2942c3c44a7 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 804.027778] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 804.028086] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a70f4163-f409-4076-8271-f4f3ceff71e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.035616] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 804.035616] env[61991]: value = "task-1129735" [ 804.035616] env[61991]: _type = "Task" [ 804.035616] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.044974] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129735, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.139138] env[61991]: DEBUG oslo_vmware.api [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129734, 'name': PowerOffVM_Task, 'duration_secs': 0.329686} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.139138] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.139138] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.139138] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49a3dfe6-c67a-4009-aaf5-06fc55146768 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.211525] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.211764] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.212063] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Deleting the datastore file [datastore2] 884e465b-e14f-4114-81a6-40e2e8694558 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.212369] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ead54d0f-e7f0-44b6-8386-191d9494770a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.219644] env[61991]: DEBUG nova.objects.instance [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lazy-loading 'numa_topology' on Instance uuid f78ef63b-453e-45d3-959b-4b0c1922b53e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.222929] env[61991]: DEBUG oslo_vmware.api [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for the task: (returnval){ [ 804.222929] env[61991]: value = "task-1129737" [ 804.222929] env[61991]: _type = "Task" [ 804.222929] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.231819] env[61991]: DEBUG oslo_vmware.api [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.272160] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d6fdeb31-b767-4477-943c-2dc242b97b24 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "0aa67bb1-59d3-42e4-8f12-47e61a130dfc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.033s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.318781] env[61991]: INFO nova.compute.manager [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Took 40.71 seconds to build instance. [ 804.389017] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129732, 'name': CreateVM_Task, 'duration_secs': 0.608614} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.389017] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.389017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.389017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.389299] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.389496] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d58b3fb0-9681-4496-937f-2104a441d81e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.400096] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 804.400096] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52559e0f-2886-ad02-eb11-dbfcabcb855f" [ 804.400096] env[61991]: _type = "Task" [ 804.400096] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.413665] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52559e0f-2886-ad02-eb11-dbfcabcb855f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.417229] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.441759] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622385} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.442160] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 62b3e5ae-b90c-47f8-95c4-14587dbf647d/62b3e5ae-b90c-47f8-95c4-14587dbf647d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.442616] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.442704] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16cc6c41-0709-4dfd-855e-ed9c1cb7d001 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.449499] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 804.449499] env[61991]: value = "task-1129738" [ 804.449499] env[61991]: _type = "Task" [ 804.449499] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.463754] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129738, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.548740] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129735, 'name': Destroy_Task, 'duration_secs': 0.413603} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.549131] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Destroyed the VM [ 804.549317] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 804.549593] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c0d1f280-8559-4ca4-8f56-cd826059d886 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.556695] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 804.556695] env[61991]: value = "task-1129739" [ 804.556695] env[61991]: _type = "Task" [ 804.556695] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.566697] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129739, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.612743] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.613139] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.613416] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.613658] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.613876] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.616411] env[61991]: INFO nova.compute.manager [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Terminating instance [ 804.618431] env[61991]: DEBUG nova.compute.manager [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.618590] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.619431] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4736131c-944d-40f4-90cd-a6a452ca7cce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.627080] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.627367] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e720619f-9fbb-440f-9fde-6fcc117c6c00 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.634517] env[61991]: DEBUG oslo_vmware.api [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 804.634517] env[61991]: value = "task-1129740" [ 804.634517] env[61991]: _type = "Task" [ 804.634517] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.643351] env[61991]: DEBUG oslo_vmware.api [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.725185] env[61991]: DEBUG nova.objects.base [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 804.738697] env[61991]: DEBUG oslo_vmware.api [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Task: {'id': task-1129737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.383586} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.738985] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 804.739178] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 804.739375] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.739618] env[61991]: INFO nova.compute.manager [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Took 1.14 seconds to destroy the instance on the hypervisor. [ 804.741552] env[61991]: DEBUG oslo.service.loopingcall [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.741552] env[61991]: DEBUG nova.compute.manager [-] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.741552] env[61991]: DEBUG nova.network.neutron [-] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 804.821654] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "671e0197-4f96-4660-be99-ea1d1c0588e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.821928] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.822192] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "671e0197-4f96-4660-be99-ea1d1c0588e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.822347] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.822516] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.824926] env[61991]: DEBUG oslo_concurrency.lockutils [None req-fe54f4b9-4909-44c0-b117-c9224aaeef75 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.953s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.825722] env[61991]: INFO nova.compute.manager [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Terminating instance [ 804.828224] env[61991]: DEBUG nova.compute.manager [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.828498] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.829402] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d314a47c-6aa1-4db5-a0dd-697e1cfbbc7e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.842422] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.845393] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a56b47a-6da6-4e12-bb83-42ba6526d1e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.853922] env[61991]: DEBUG oslo_vmware.api [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 804.853922] env[61991]: value = "task-1129741" [ 804.853922] env[61991]: _type = "Task" [ 804.853922] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.863899] env[61991]: DEBUG oslo_vmware.api [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129741, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.895243] env[61991]: DEBUG nova.network.neutron [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Updating instance_info_cache with network_info: [{"id": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "address": "fa:16:3e:ee:36:73", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9be477d-86", "ovs_interfaceid": "c9be477d-8647-4664-8ebf-1f1fe5e45492", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.912425] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52559e0f-2886-ad02-eb11-dbfcabcb855f, 'name': SearchDatastore_Task, 'duration_secs': 0.017381} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.916820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.916820] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.916820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.916820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.917160] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.921970] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76f391a5-b06b-4260-bf9b-162da745bd54 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.926872] env[61991]: DEBUG oslo_vmware.api [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129719, 'name': ReconfigVM_Task, 'duration_secs': 5.833851} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.927463] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.927463] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Reconfigured VM to detach interface {{(pid=61991) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 804.940175] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.940175] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.940175] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9189929-8754-429b-b28e-b55753d7b3b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.946301] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 804.946301] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52935ab1-50c8-e2ba-77ff-a18851db29e7" [ 804.946301] env[61991]: _type = "Task" [ 804.946301] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.963147] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52935ab1-50c8-e2ba-77ff-a18851db29e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.970691] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129738, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137196} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.974507] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.975843] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43dcf58-376f-4ff0-a110-e0ebe23588ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.002420] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 62b3e5ae-b90c-47f8-95c4-14587dbf647d/62b3e5ae-b90c-47f8-95c4-14587dbf647d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.006986] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c5d09df-552d-457b-a90e-e1a71663b9cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.027254] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 805.027254] env[61991]: value = "task-1129742" [ 805.027254] env[61991]: _type = "Task" [ 805.027254] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.036781] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.073052] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129739, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.089115] env[61991]: DEBUG nova.network.neutron [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updated VIF entry in instance network info cache for port 9452191d-e715-4ad9-a48e-43aebb90b551. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 805.089115] env[61991]: DEBUG nova.network.neutron [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.123052] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e742bd-5560-4789-8cb0-31f5b2e680b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.132642] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1d1508-5c4d-438c-a39c-4080496adee8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.173436] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5868f17-d47a-4758-b0dc-49dc1fad3962 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.176760] env[61991]: DEBUG oslo_vmware.api [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129740, 'name': PowerOffVM_Task, 'duration_secs': 0.264856} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.177051] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.177237] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.177886] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f363773c-7719-43fe-8944-a2bc47c22d80 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.183270] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a56c46-a69d-4edd-9133-cc6e508fb4ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.200021] env[61991]: DEBUG nova.compute.provider_tree [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.243562] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.243798] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.243984] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleting the datastore file [datastore1] 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.244271] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4889822-34d0-4f28-a23f-63b0eb391f35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.253156] env[61991]: DEBUG oslo_vmware.api [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 805.253156] env[61991]: value = "task-1129744" [ 805.253156] env[61991]: _type = "Task" [ 805.253156] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.260433] env[61991]: DEBUG oslo_vmware.api [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.365103] env[61991]: DEBUG oslo_vmware.api [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129741, 'name': PowerOffVM_Task, 'duration_secs': 0.331855} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.365103] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.365768] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.365768] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0585f2c8-e803-45ed-8dd1-33de5f08e37f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.401396] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "refresh_cache-f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.426289] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.426571] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.426788] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleting the datastore file [datastore2] 671e0197-4f96-4660-be99-ea1d1c0588e1 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.429785] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08f094d8-18a6-4391-a014-ff9c5a161f18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.441909] env[61991]: DEBUG oslo_vmware.api [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for the task: (returnval){ [ 805.441909] env[61991]: value = "task-1129746" [ 805.441909] env[61991]: _type = "Task" [ 805.441909] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.456427] env[61991]: DEBUG oslo_vmware.api [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.463838] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52935ab1-50c8-e2ba-77ff-a18851db29e7, 'name': SearchDatastore_Task, 'duration_secs': 0.02241} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.465608] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-580b1684-bfe6-44e7-9f1d-71e0a8173010 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.475107] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 805.475107] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522185e1-a95f-b029-80d8-f80d85f1b552" [ 805.475107] env[61991]: _type = "Task" [ 805.475107] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.486172] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522185e1-a95f-b029-80d8-f80d85f1b552, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.542309] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129742, 'name': ReconfigVM_Task, 'duration_secs': 0.482734} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.542309] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 62b3e5ae-b90c-47f8-95c4-14587dbf647d/62b3e5ae-b90c-47f8-95c4-14587dbf647d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.542309] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4520a6c-817f-4150-a154-f876005771d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.551530] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 805.551530] env[61991]: value = "task-1129747" [ 805.551530] env[61991]: _type = "Task" [ 805.551530] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.566516] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129747, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.574911] env[61991]: DEBUG oslo_vmware.api [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129739, 'name': RemoveSnapshot_Task, 'duration_secs': 0.909104} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.575262] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 805.592230] env[61991]: DEBUG oslo_concurrency.lockutils [req-d5001962-1a40-48a0-83bc-3eaa2a4d4cb9 req-935f0772-c9df-4318-99a1-f76105eab158 service nova] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.729014] env[61991]: ERROR nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [req-c0954e16-ee4b-40ab-af63-afdb1c3788c2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c0954e16-ee4b-40ab-af63-afdb1c3788c2"}]} [ 805.750797] env[61991]: DEBUG nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 805.763730] env[61991]: DEBUG oslo_vmware.api [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323812} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.764858] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.764858] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.764858] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.765199] env[61991]: INFO nova.compute.manager [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Took 1.15 seconds to destroy the instance on the hypervisor. [ 805.765463] env[61991]: DEBUG oslo.service.loopingcall [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.765729] env[61991]: DEBUG nova.compute.manager [-] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.765924] env[61991]: DEBUG nova.network.neutron [-] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 805.776047] env[61991]: DEBUG nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 805.776047] env[61991]: DEBUG nova.compute.provider_tree [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.786690] env[61991]: DEBUG nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 805.801912] env[61991]: DEBUG nova.network.neutron [-] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.812181] env[61991]: DEBUG nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 805.951297] env[61991]: DEBUG oslo_vmware.api [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Task: {'id': task-1129746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287093} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.951297] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.951569] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.951617] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.951796] env[61991]: INFO nova.compute.manager [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 805.952078] env[61991]: DEBUG oslo.service.loopingcall [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.955330] env[61991]: DEBUG nova.compute.manager [-] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.955469] env[61991]: DEBUG nova.network.neutron [-] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 805.961328] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.961954] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fabb691f-495e-41e8-8cb1-2112eb66ddba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.968913] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 805.968913] env[61991]: value = "task-1129748" [ 805.968913] env[61991]: _type = "Task" [ 805.968913] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.978018] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129748, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.993642] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522185e1-a95f-b029-80d8-f80d85f1b552, 'name': SearchDatastore_Task, 'duration_secs': 0.026749} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.993916] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.994238] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e/d36bd981-9ea3-46f6-8376-ac1e0c3bf61e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.994501] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23d1d265-cfc0-4fc3-8c5f-a4d3d646759e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.003580] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 806.003580] env[61991]: value = "task-1129749" [ 806.003580] env[61991]: _type = "Task" [ 806.003580] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.012825] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129749, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.062270] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129747, 'name': Rename_Task, 'duration_secs': 0.183471} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.062688] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.062903] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a46baaf-8f23-491c-b8e3-8a5d5ae4b0d4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.078352] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 806.078352] env[61991]: value = "task-1129750" [ 806.078352] env[61991]: _type = "Task" [ 806.078352] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.086548] env[61991]: WARNING nova.compute.manager [None req-09cce3dd-8edc-43b4-9492-3505cad946c5 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Image not found during snapshot: nova.exception.ImageNotFound: Image e5b4302d-ea69-40e0-9557-d2942c3c44a7 could not be found. [ 806.097541] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129750, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.103184] env[61991]: DEBUG nova.compute.manager [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Received event network-vif-deleted-50b6a76d-9313-4b7c-b577-09c086277647 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.103358] env[61991]: DEBUG nova.compute.manager [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-vif-deleted-a20e94c9-8ee3-4696-b020-423d80f00dc9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.103529] env[61991]: INFO nova.compute.manager [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Neutron deleted interface a20e94c9-8ee3-4696-b020-423d80f00dc9; detaching it from the instance and deleting it from the info cache [ 806.104150] env[61991]: DEBUG nova.network.neutron [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e08c38be-49fa-4a29-9cb1-37e15930e258", "address": "fa:16:3e:41:e8:b9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape08c38be-49", "ovs_interfaceid": "e08c38be-49fa-4a29-9cb1-37e15930e258", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.169982] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9ecd74-56ef-49b6-ad61-4f263eedd0e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.178610] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dade4ecd-8ed1-4891-b4f3-75a75f5be1f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.216096] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88fb7f8-a076-4b58-b201-2d1faef8bdc7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.220271] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "49324fef-ad48-451b-a5ce-d9a1231137db" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.220535] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.220724] env[61991]: DEBUG nova.compute.manager [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.221618] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc722043-3ee6-4baf-9fbd-a5b6fe77640c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.235538] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222354cb-2a98-48d9-8e0c-c21c0c6ea1b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.241280] env[61991]: DEBUG nova.compute.manager [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 806.241995] env[61991]: DEBUG nova.objects.instance [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'flavor' on Instance uuid 49324fef-ad48-451b-a5ce-d9a1231137db {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 806.256028] env[61991]: DEBUG nova.compute.provider_tree [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.306661] env[61991]: INFO nova.compute.manager [-] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Took 1.57 seconds to deallocate network for instance. [ 806.375310] env[61991]: DEBUG nova.compute.manager [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-vif-deleted-e08c38be-49fa-4a29-9cb1-37e15930e258 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.376099] env[61991]: INFO nova.compute.manager [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Neutron deleted interface e08c38be-49fa-4a29-9cb1-37e15930e258; detaching it from the instance and deleting it from the info cache [ 806.376099] env[61991]: DEBUG nova.network.neutron [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.412404] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.412635] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.412857] env[61991]: DEBUG nova.network.neutron [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.480619] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129748, 'name': PowerOffVM_Task, 'duration_secs': 0.250324} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.480971] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 806.482160] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b18905b-15b0-4deb-a586-ca8be8830cba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.505278] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee53285-a7d6-4381-8b17-0b62ab6d65c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.529779] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129749, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.557516] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 806.558526] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-308002a1-7aaf-4395-835b-9d5e56cd853f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.565722] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 806.565722] env[61991]: value = "task-1129751" [ 806.565722] env[61991]: _type = "Task" [ 806.565722] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.576365] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 806.576672] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.577018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.577208] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.577403] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.577723] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5920dd0-e857-4aa6-b890-c45aa79992df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.594311] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129750, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.597096] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.598221] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.598221] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc0db17b-12d9-4f49-a73a-45bcb68c3bb0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.604160] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 806.604160] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c3401b-20d3-f2aa-954e-2d9de273ae97" [ 806.604160] env[61991]: _type = "Task" [ 806.604160] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.608581] env[61991]: DEBUG oslo_concurrency.lockutils [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.608581] env[61991]: DEBUG oslo_concurrency.lockutils [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] Acquired lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.609133] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a968c7-0c43-495d-a13c-d901eecd7b7e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.621118] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c3401b-20d3-f2aa-954e-2d9de273ae97, 'name': SearchDatastore_Task, 'duration_secs': 0.013883} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.634827] env[61991]: DEBUG oslo_concurrency.lockutils [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] Releasing lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.635179] env[61991]: WARNING nova.compute.manager [req-c75170cc-a889-4344-af58-2f64557433aa req-05c74c44-aac0-4828-b11b-34ace8f5d893 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Detach interface failed, port_id=a20e94c9-8ee3-4696-b020-423d80f00dc9, reason: No device with interface-id a20e94c9-8ee3-4696-b020-423d80f00dc9 exists on VM: nova.exception.NotFound: No device with interface-id a20e94c9-8ee3-4696-b020-423d80f00dc9 exists on VM [ 806.635713] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f106656f-8424-442f-a1c8-ca61b78fac84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.642793] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 806.642793] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526567de-f4b2-5d66-2b89-a9c16a51f766" [ 806.642793] env[61991]: _type = "Task" [ 806.642793] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.651850] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526567de-f4b2-5d66-2b89-a9c16a51f766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.749210] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 806.749592] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64e55522-2c6d-4bb0-ace9-7b8e3814f1bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.757246] env[61991]: DEBUG oslo_vmware.api [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 806.757246] env[61991]: value = "task-1129752" [ 806.757246] env[61991]: _type = "Task" [ 806.757246] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.761891] env[61991]: DEBUG nova.scheduler.client.report [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.768897] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.768897] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "80d05278-b516-4408-94b0-11bc93500b5c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.768897] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.768897] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.768897] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "80d05278-b516-4408-94b0-11bc93500b5c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.769814] env[61991]: INFO nova.compute.manager [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Terminating instance [ 806.775766] env[61991]: DEBUG nova.network.neutron [-] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.779038] env[61991]: DEBUG oslo_vmware.api [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.779038] env[61991]: DEBUG nova.compute.manager [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.779038] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.779038] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50cec06-2b18-4018-b29d-5e79f21b7345 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.787507] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 806.787649] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c9c2834-6343-4212-8b1b-5998175a0061 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.795196] env[61991]: DEBUG oslo_vmware.api [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 806.795196] env[61991]: value = "task-1129753" [ 806.795196] env[61991]: _type = "Task" [ 806.795196] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.804973] env[61991]: DEBUG oslo_vmware.api [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.816050] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.879265] env[61991]: DEBUG oslo_concurrency.lockutils [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] Acquiring lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.013618] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129749, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621283} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.013903] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e/d36bd981-9ea3-46f6-8376-ac1e0c3bf61e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.014151] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.014404] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fafc9e5-9224-4a1e-b950-a92623380760 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.021093] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 807.021093] env[61991]: value = "task-1129754" [ 807.021093] env[61991]: _type = "Task" [ 807.021093] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.029127] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.095258] env[61991]: DEBUG oslo_vmware.api [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129750, 'name': PowerOnVM_Task, 'duration_secs': 0.642956} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.097940] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.097940] env[61991]: INFO nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Took 9.95 seconds to spawn the instance on the hypervisor. [ 807.097940] env[61991]: DEBUG nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.098342] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a983b4b-63f1-461f-9f54-3622fafd561f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.156811] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526567de-f4b2-5d66-2b89-a9c16a51f766, 'name': SearchDatastore_Task, 'duration_secs': 0.016578} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.157588] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.157829] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. {{(pid=61991) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 807.158084] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-678253d0-16f9-4efc-bc53-7d37caf8ecff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.167724] env[61991]: DEBUG nova.network.neutron [-] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.171226] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 807.171226] env[61991]: value = "task-1129755" [ 807.171226] env[61991]: _type = "Task" [ 807.171226] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.186236] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.270059] env[61991]: DEBUG oslo_vmware.api [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129752, 'name': PowerOffVM_Task, 'duration_secs': 0.197721} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.270407] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 807.270991] env[61991]: DEBUG nova.compute.manager [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.271818] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.557s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.275314] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8838db06-5971-43de-aa89-584e9a00ff07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.278591] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.941s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.278838] env[61991]: DEBUG nova.objects.instance [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lazy-loading 'resources' on Instance uuid c279a82f-bf4c-4a59-b04a-38d31070b0e5 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.283028] env[61991]: INFO nova.compute.manager [-] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Took 1.52 seconds to deallocate network for instance. [ 807.308190] env[61991]: DEBUG oslo_vmware.api [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129753, 'name': PowerOffVM_Task, 'duration_secs': 0.212788} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.308190] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 807.308190] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 807.308190] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed5ba0d1-891d-4a05-8390-af58eabe32e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.434489] env[61991]: DEBUG nova.network.neutron [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [{"id": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "address": "fa:16:3e:1d:32:ef", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea66ec35-3e", "ovs_interfaceid": "ea66ec35-3e8b-40d2-baa8-a6d811b6e7be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.444900] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 807.445173] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 807.445360] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleting the datastore file [datastore1] 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.445622] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e9c5a75-7ee0-433c-a505-f7f9a6921b1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.456026] env[61991]: DEBUG oslo_vmware.api [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 807.456026] env[61991]: value = "task-1129757" [ 807.456026] env[61991]: _type = "Task" [ 807.456026] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.459485] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "3e62f219-7fec-470f-9d0a-e292d62e076c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.459895] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.460179] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "3e62f219-7fec-470f-9d0a-e292d62e076c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.460420] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.460632] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.465879] env[61991]: DEBUG oslo_vmware.api [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129757, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.466461] env[61991]: INFO nova.compute.manager [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Terminating instance [ 807.468611] env[61991]: DEBUG nova.compute.manager [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.468857] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.469736] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d81cc2-529a-449c-b9e4-ebdfbf3a7df6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.478306] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.478677] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1fff26f-08ea-4868-9069-1c36af6b2938 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.487603] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 807.487603] env[61991]: value = "task-1129758" [ 807.487603] env[61991]: _type = "Task" [ 807.487603] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.496174] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.533026] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120268} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.533026] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.533731] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25de2ad9-ae39-4da4-850c-463decce2bfb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.566331] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e/d36bd981-9ea3-46f6-8376-ac1e0c3bf61e.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.566331] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad5fe138-3f5d-497f-afe6-2487cad619f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.584402] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 807.584402] env[61991]: value = "task-1129759" [ 807.584402] env[61991]: _type = "Task" [ 807.584402] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.595041] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129759, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.618911] env[61991]: INFO nova.compute.manager [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Took 40.32 seconds to build instance. [ 807.684386] env[61991]: INFO nova.compute.manager [-] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Took 1.73 seconds to deallocate network for instance. [ 807.697618] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129755, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.785913] env[61991]: DEBUG oslo_concurrency.lockutils [None req-01ed68fd-8a86-4940-bd5d-0006677db256 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 42.909s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.789783] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 14.317s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.789783] env[61991]: INFO nova.compute.manager [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Unshelving [ 807.790864] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.796170] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a7a77d0-bccb-44e5-a9e7-5757fb90dd88 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.575s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.939068] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.968141] env[61991]: DEBUG oslo_vmware.api [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129757, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.404698} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.968687] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 807.969029] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 807.969685] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.969874] env[61991]: INFO nova.compute.manager [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Took 1.19 seconds to destroy the instance on the hypervisor. [ 807.972023] env[61991]: DEBUG oslo.service.loopingcall [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.972023] env[61991]: DEBUG nova.compute.manager [-] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.972023] env[61991]: DEBUG nova.network.neutron [-] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.003027] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.028589] env[61991]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port e08c38be-49fa-4a29-9cb1-37e15930e258 could not be found.", "detail": ""}} {{(pid=61991) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 808.028874] env[61991]: DEBUG nova.network.neutron [-] Unable to show port e08c38be-49fa-4a29-9cb1-37e15930e258 as it no longer exists. {{(pid=61991) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 808.106103] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129759, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.111149] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d85189-dd71-4f2e-bb1c-1f444c6b895a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.119891] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9538378-9964-48fb-8c36-4c8eb2fbc69d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.123373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ff9037bc-10c8-47cf-8c6f-291963010443 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.575s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.154012] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e02e7e7-b94e-4b38-ba3d-c0a5d378ce42 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.166719] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a16016-6f3f-47b7-98e7-85af8762fe60 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.184300] env[61991]: DEBUG nova.compute.provider_tree [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.197574] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.197966] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129755, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59956} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.198778] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. [ 808.199948] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc50be22-ea4e-4a46-aa19-2c760867b006 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.234703] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.234703] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dde68236-4aa9-46d3-8688-066e27a98f79 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.251535] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 808.251535] env[61991]: value = "task-1129760" [ 808.251535] env[61991]: _type = "Task" [ 808.251535] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.254112] env[61991]: DEBUG nova.compute.manager [req-cde006c9-5f92-4edb-9df9-bffa0c7e8fe2 req-e2503b6d-1ef6-4e03-9902-62d6faed781c service nova] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Received event network-vif-deleted-790d5666-4e4c-4dcf-8f02-6a40df7a7666 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.255965] env[61991]: DEBUG nova.compute.manager [req-cde006c9-5f92-4edb-9df9-bffa0c7e8fe2 req-e2503b6d-1ef6-4e03-9902-62d6faed781c service nova] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Received event network-vif-deleted-35afc727-b479-4f5e-b793-9e8be36b9975 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.264625] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129760, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.442487] env[61991]: DEBUG oslo_concurrency.lockutils [None req-92cc6f95-3743-42d1-a975-dff3420ec781 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-80d05278-b516-4408-94b0-11bc93500b5c-a20e94c9-8ee3-4696-b020-423d80f00dc9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.143s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.494922] env[61991]: DEBUG nova.compute.manager [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Received event network-changed-f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.495147] env[61991]: DEBUG nova.compute.manager [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Refreshing instance network info cache due to event network-changed-f9418974-50f5-48bc-a8cc-dd016b21728a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.495726] env[61991]: DEBUG oslo_concurrency.lockutils [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] Acquiring lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.495726] env[61991]: DEBUG oslo_concurrency.lockutils [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] Acquired lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.495726] env[61991]: DEBUG nova.network.neutron [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Refreshing network info cache for port f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.505837] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129758, 'name': PowerOffVM_Task, 'duration_secs': 0.965363} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.506616] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.506799] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.507146] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ecfdf81a-cb78-47ba-9a6f-43ea85e631fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.589216] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.589539] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.589749] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleting the datastore file [datastore1] 3e62f219-7fec-470f-9d0a-e292d62e076c {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.590067] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f70d3115-11ef-4e16-b772-3f995d6dc9bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.596933] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 808.596933] env[61991]: value = "task-1129762" [ 808.596933] env[61991]: _type = "Task" [ 808.596933] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.608979] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129759, 'name': ReconfigVM_Task, 'duration_secs': 0.697972} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.612324] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Reconfigured VM instance instance-00000046 to attach disk [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e/d36bd981-9ea3-46f6-8376-ac1e0c3bf61e.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.613085] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129762, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.613579] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea5a3983-6ddd-4a44-b23c-aab1653bded2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.619554] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 808.619554] env[61991]: value = "task-1129763" [ 808.619554] env[61991]: _type = "Task" [ 808.619554] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.628850] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129763, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.694941] env[61991]: DEBUG nova.scheduler.client.report [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.776938] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129760, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.825941] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.867373] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5218d801-4ab0-b8cf-8b00-ba41e7d92309/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 808.868657] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ab45d8-84df-4e4a-967d-f3af94bfe1ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.874838] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5218d801-4ab0-b8cf-8b00-ba41e7d92309/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 808.875037] env[61991]: ERROR oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5218d801-4ab0-b8cf-8b00-ba41e7d92309/disk-0.vmdk due to incomplete transfer. [ 808.875282] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-45fff2c9-dfe7-419d-a5e3-ea725921c839 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.883507] env[61991]: DEBUG oslo_vmware.rw_handles [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5218d801-4ab0-b8cf-8b00-ba41e7d92309/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 808.883728] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Uploaded image add4fbc4-635d-4eaf-945a-ac91215c88f3 to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 808.888862] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 808.889264] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e3270358-1c87-4241-930d-181c92c4e059 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.895954] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 808.895954] env[61991]: value = "task-1129764" [ 808.895954] env[61991]: _type = "Task" [ 808.895954] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.904957] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129764, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.110685] env[61991]: DEBUG oslo_vmware.api [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129762, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229028} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.113129] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.113129] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 809.113129] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.113129] env[61991]: INFO nova.compute.manager [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Took 1.64 seconds to destroy the instance on the hypervisor. [ 809.113129] env[61991]: DEBUG oslo.service.loopingcall [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.113129] env[61991]: DEBUG nova.compute.manager [-] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.113129] env[61991]: DEBUG nova.network.neutron [-] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 809.129509] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129763, 'name': Rename_Task, 'duration_secs': 0.177444} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.129820] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.130110] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c46fd190-c47a-4494-8c00-f686e0cd68c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.136050] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 809.136050] env[61991]: value = "task-1129765" [ 809.136050] env[61991]: _type = "Task" [ 809.136050] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.143839] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.197528] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.919s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.200384] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.091s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.200713] env[61991]: DEBUG nova.objects.instance [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lazy-loading 'resources' on Instance uuid 700ce4a7-d597-449a-9379-0cfb0c8f82fc {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.220932] env[61991]: INFO nova.scheduler.client.report [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Deleted allocations for instance c279a82f-bf4c-4a59-b04a-38d31070b0e5 [ 809.268295] env[61991]: INFO nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Rebuilding instance [ 809.272167] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129760, 'name': ReconfigVM_Task, 'duration_secs': 0.528091} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.273526] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Reconfigured VM instance instance-00000043 to attach disk [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.274568] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bc01ab-0c5b-444a-ba92-df46b0744cf0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.314451] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e73fff51-1b09-4db2-b06e-e741a64b4243 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.331756] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 809.331756] env[61991]: value = "task-1129766" [ 809.331756] env[61991]: _type = "Task" [ 809.331756] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.345949] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129766, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.355262] env[61991]: DEBUG nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.355262] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b086bef-428d-4279-93e5-ef9f3ae5d085 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.407954] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129764, 'name': Destroy_Task, 'duration_secs': 0.333467} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.407954] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroyed the VM [ 809.407954] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 809.407954] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d22434ee-8ab5-4462-85b1-15f489fa580e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.416416] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 809.416416] env[61991]: value = "task-1129767" [ 809.416416] env[61991]: _type = "Task" [ 809.416416] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.427185] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129767, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.541146] env[61991]: DEBUG nova.network.neutron [-] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.647756] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129765, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.735230] env[61991]: DEBUG oslo_concurrency.lockutils [None req-169cac3b-e10f-400c-ab6f-adde1b6f6946 tempest-MultipleCreateTestJSON-137875419 tempest-MultipleCreateTestJSON-137875419-project-member] Lock "c279a82f-bf4c-4a59-b04a-38d31070b0e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.907s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.738310] env[61991]: DEBUG nova.network.neutron [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Updated VIF entry in instance network info cache for port f9418974-50f5-48bc-a8cc-dd016b21728a. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.738310] env[61991]: DEBUG nova.network.neutron [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Updating instance_info_cache with network_info: [{"id": "f9418974-50f5-48bc-a8cc-dd016b21728a", "address": "fa:16:3e:f8:d8:1a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9418974-50", "ovs_interfaceid": "f9418974-50f5-48bc-a8cc-dd016b21728a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.847671] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129766, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.866162] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.866162] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54a7ef3f-2e72-490e-a207-ab0e0d04a35d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.875332] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 809.875332] env[61991]: value = "task-1129768" [ 809.875332] env[61991]: _type = "Task" [ 809.875332] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.889058] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 809.889058] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.890444] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2bb014-20f1-455b-8f00-e5bb4dfae6aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.898667] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.898972] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef1523d8-2c26-4b8d-8d88-f08239e65e64 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.929427] env[61991]: DEBUG oslo_vmware.api [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129767, 'name': RemoveSnapshot_Task, 'duration_secs': 0.509243} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.929683] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 809.929920] env[61991]: INFO nova.compute.manager [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 15.76 seconds to snapshot the instance on the hypervisor. [ 809.988320] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 809.988857] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 809.989381] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore2] 49324fef-ad48-451b-a5ce-d9a1231137db {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.993791] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfa9c7db-dc52-449f-a111-73b57c0e4538 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.005022] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 810.005022] env[61991]: value = "task-1129770" [ 810.005022] env[61991]: _type = "Task" [ 810.005022] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.012260] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.045410] env[61991]: INFO nova.compute.manager [-] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Took 2.07 seconds to deallocate network for instance. [ 810.046902] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e4b283-6ae0-4434-b712-21a6387cac03 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.060501] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94381e90-71ec-4fd2-a0b7-bf3a6deb2964 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.099727] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a572084f-aa0a-4be9-b21b-db43fef01603 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.108115] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84591b22-8ff7-4a14-b42e-0c74dca85d0f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.123637] env[61991]: DEBUG nova.compute.provider_tree [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.147818] env[61991]: DEBUG oslo_vmware.api [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129765, 'name': PowerOnVM_Task, 'duration_secs': 0.919994} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.148608] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.148898] env[61991]: INFO nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Took 10.40 seconds to spawn the instance on the hypervisor. [ 810.149112] env[61991]: DEBUG nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.149964] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a22b63-d2d3-45c9-836c-10dc62ef2872 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.244806] env[61991]: DEBUG oslo_concurrency.lockutils [req-2b179cf6-9350-47b6-996c-4b17df109a5a req-d387ad14-2dfe-47d2-ae9c-ac8b89f5d797 service nova] Releasing lock "refresh_cache-62b3e5ae-b90c-47f8-95c4-14587dbf647d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.244806] env[61991]: DEBUG nova.network.neutron [-] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.294016] env[61991]: DEBUG nova.compute.manager [req-2036a118-15fd-48e0-b72a-d1456694d60f req-26faa251-cf2d-474d-a03b-f3baab0ccfdd service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Received event network-vif-deleted-ea66ec35-3e8b-40d2-baa8-a6d811b6e7be {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.294237] env[61991]: DEBUG nova.compute.manager [req-2036a118-15fd-48e0-b72a-d1456694d60f req-26faa251-cf2d-474d-a03b-f3baab0ccfdd service nova] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Received event network-vif-deleted-3ea67247-a11c-4dfb-8875-ffc387e7678e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.347245] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129766, 'name': ReconfigVM_Task, 'duration_secs': 0.517619} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.347245] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.347245] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4877cccc-cc33-4a2d-bfaa-aed9ddcef707 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.353583] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 810.353583] env[61991]: value = "task-1129771" [ 810.353583] env[61991]: _type = "Task" [ 810.353583] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.361126] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129771, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.485420] env[61991]: DEBUG nova.compute.manager [None req-e41ce0ae-702f-4082-a7a7-0c4f957a8b58 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Found 2 images (rotation: 2) {{(pid=61991) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 810.513330] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.459608} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.513595] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.513787] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.513975] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.558955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.627660] env[61991]: DEBUG nova.scheduler.client.report [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.669845] env[61991]: INFO nova.compute.manager [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Took 39.40 seconds to build instance. [ 810.751020] env[61991]: INFO nova.compute.manager [-] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Took 1.64 seconds to deallocate network for instance. [ 810.864320] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129771, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.132520] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.932s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.134942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.804s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.137031] env[61991]: INFO nova.compute.claims [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.157171] env[61991]: INFO nova.scheduler.client.report [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Deleted allocations for instance 700ce4a7-d597-449a-9379-0cfb0c8f82fc [ 811.172107] env[61991]: DEBUG oslo_concurrency.lockutils [None req-809a22b6-cd3a-43ed-80e7-3e676ccdbbe2 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.543s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.254971] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.363139] env[61991]: DEBUG oslo_vmware.api [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129771, 'name': PowerOnVM_Task, 'duration_secs': 0.675747} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.363362] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.366199] env[61991]: DEBUG nova.compute.manager [None req-8301f59e-1694-4279-b09b-dd736e2a79cc tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.366951] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f144d0-80cf-427c-a712-a90a25eb0ea6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.544768] env[61991]: DEBUG nova.compute.manager [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.545749] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df19e139-f894-4faf-8310-d8cfbd038e7d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.550697] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.550960] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.551438] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.551692] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.551857] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.552114] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.552348] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.552512] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.552691] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.552886] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.553144] env[61991]: DEBUG nova.virt.hardware [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.553993] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd5bec9-de49-4ee3-89d4-7a87ce5dbcb5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.565204] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ec5909-7ff0-4811-84bf-7e6661d1835c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.579383] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:c3:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa5c2434-a034-4ee0-b847-7108e636ef8d', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.587585] env[61991]: DEBUG oslo.service.loopingcall [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.588654] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.588875] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c63621f1-27eb-478a-97e2-5bc3eab19020 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.609649] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.609649] env[61991]: value = "task-1129772" [ 811.609649] env[61991]: _type = "Task" [ 811.609649] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.618886] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129772, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.666225] env[61991]: DEBUG oslo_concurrency.lockutils [None req-073fb77c-04f1-4045-8ca2-8006acf56c8b tempest-ServersTestJSON-764805758 tempest-ServersTestJSON-764805758-project-member] Lock "700ce4a7-d597-449a-9379-0cfb0c8f82fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.551s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.066063] env[61991]: INFO nova.compute.manager [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] instance snapshotting [ 812.066716] env[61991]: DEBUG nova.objects.instance [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'flavor' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.124383] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129772, 'name': CreateVM_Task, 'duration_secs': 0.303205} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.124437] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.125155] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.125320] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.125638] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.125893] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-774bd3f9-c297-4aef-870b-49c80e8aac7d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.131714] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 812.131714] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52547ca6-fdf7-5bdd-d658-c4222c603950" [ 812.131714] env[61991]: _type = "Task" [ 812.131714] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.139964] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52547ca6-fdf7-5bdd-d658-c4222c603950, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.203782] env[61991]: DEBUG nova.compute.manager [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Received event network-changed-9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.203988] env[61991]: DEBUG nova.compute.manager [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Refreshing instance network info cache due to event network-changed-9452191d-e715-4ad9-a48e-43aebb90b551. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.204240] env[61991]: DEBUG oslo_concurrency.lockutils [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.204547] env[61991]: DEBUG oslo_concurrency.lockutils [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.204611] env[61991]: DEBUG nova.network.neutron [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Refreshing network info cache for port 9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 812.404633] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cdb384-a774-4263-900b-6daa72ec0f4c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.413300] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f904d2cd-ee26-4263-a97a-171b53c1792c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.452603] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae7729a-d4f3-4090-a5e7-fc3d663dd060 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.464553] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c922dbea-f21b-4365-b2fe-8990808955b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.480488] env[61991]: DEBUG nova.compute.provider_tree [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.576937] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe80539-72cb-4f51-ab5e-87c6342c5c3c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.601813] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c043c8c-49f5-4f6e-be7e-ac6890eecbe8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.646951] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52547ca6-fdf7-5bdd-d658-c4222c603950, 'name': SearchDatastore_Task, 'duration_secs': 0.027882} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.646951] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.646951] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.646951] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.646951] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.646951] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.647339] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e9c98c8-76e7-44ca-9614-a21fa55f1da4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.656252] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.656252] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.660020] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abb91364-f302-4b1d-837e-804a98d50a8f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.665340] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 812.665340] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5238c646-aeb2-d22e-1d20-ed2abd4dc326" [ 812.665340] env[61991]: _type = "Task" [ 812.665340] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.676483] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5238c646-aeb2-d22e-1d20-ed2abd4dc326, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.987244] env[61991]: DEBUG nova.scheduler.client.report [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.086780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "96627657-d3d9-4a64-b83e-87323bd10c03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.086780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "96627657-d3d9-4a64-b83e-87323bd10c03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.104993] env[61991]: DEBUG nova.network.neutron [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updated VIF entry in instance network info cache for port 9452191d-e715-4ad9-a48e-43aebb90b551. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 813.105212] env[61991]: DEBUG nova.network.neutron [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.114991] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 813.115538] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-222a8734-875c-4cb2-a106-0c0f164f3ef7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.128213] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 813.128213] env[61991]: value = "task-1129773" [ 813.128213] env[61991]: _type = "Task" [ 813.128213] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.136086] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129773, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.156471] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.156724] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.176033] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5238c646-aeb2-d22e-1d20-ed2abd4dc326, 'name': SearchDatastore_Task, 'duration_secs': 0.010413} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.176640] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be09a6c7-1a99-4be5-9a1e-154fa7f43f80 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.184199] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 813.184199] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5290f24a-1ba2-3c7a-679e-24348708d963" [ 813.184199] env[61991]: _type = "Task" [ 813.184199] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.191730] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5290f24a-1ba2-3c7a-679e-24348708d963, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.495051] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.495582] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.498242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.031s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.498441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.500497] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.825s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.500768] env[61991]: DEBUG nova.objects.instance [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lazy-loading 'resources' on Instance uuid e3e5ec72-eec1-4155-bab2-038e3d5e38a8 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.532671] env[61991]: INFO nova.scheduler.client.report [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Deleted allocations for instance c14d99a3-950d-44d0-b330-3031a1f8a2be [ 813.592818] env[61991]: DEBUG nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.610139] env[61991]: DEBUG oslo_concurrency.lockutils [req-db57ce09-14fd-4b47-b106-f2f567258ac7 req-a27b21eb-b739-4154-9d4b-16b484a7bd0f service nova] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.636734] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129773, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.658567] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.700561] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5290f24a-1ba2-3c7a-679e-24348708d963, 'name': SearchDatastore_Task, 'duration_secs': 0.009482} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.700946] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.701247] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.701526] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-169f6ea6-b4e8-4a13-9017-cc5770ff16e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.708312] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 813.708312] env[61991]: value = "task-1129774" [ 813.708312] env[61991]: _type = "Task" [ 813.708312] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.718682] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.007916] env[61991]: DEBUG nova.compute.utils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.017171] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.017488] env[61991]: DEBUG nova.network.neutron [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 814.044830] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2da031ba-bfed-487d-a544-77f0a85171df tempest-MigrationsAdminTest-935799036 tempest-MigrationsAdminTest-935799036-project-member] Lock "c14d99a3-950d-44d0-b330-3031a1f8a2be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.108s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.104604] env[61991]: DEBUG nova.policy [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e6810d3aced4a44bbfe9ba49270bc35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a18b2a73f7643e1b89c9660a770e9da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 814.128030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.144096] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129773, 'name': CreateSnapshot_Task, 'duration_secs': 0.794144} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.144439] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 814.150028] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66cb9db7-04cc-4ae8-8f64-8774dc3e57cf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.192483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.224180] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508402} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.225307] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.225307] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.225307] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4dd52b69-7ee1-4bd8-bf81-b80727e93212 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.233039] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 814.233039] env[61991]: value = "task-1129775" [ 814.233039] env[61991]: _type = "Task" [ 814.233039] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.241091] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.351328] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abd4f9d-7ac4-4620-a35a-29a78b0f1ce3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.361324] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca327df-6abd-4e82-a45f-9c60b1b7c421 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.406834] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415197bc-e9f8-43a0-aec7-0ad91c5e0807 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.415610] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c58e58b-2aa2-450a-8b02-160dc54a64dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.431667] env[61991]: DEBUG nova.compute.provider_tree [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.518301] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.635102] env[61991]: DEBUG nova.network.neutron [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Successfully created port: d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.679406] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 814.679492] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4ad3848f-41a9-492f-92ad-4af166f4dac0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.689400] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 814.689400] env[61991]: value = "task-1129776" [ 814.689400] env[61991]: _type = "Task" [ 814.689400] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.702662] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129776, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.744965] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123193} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.749019] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.749019] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8296cf1-4802-44d4-9923-2b67494e20b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.773674] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.773993] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a4f9c86-3cda-457c-b3f8-779a06dad0c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.796886] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 814.796886] env[61991]: value = "task-1129777" [ 814.796886] env[61991]: _type = "Task" [ 814.796886] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.810782] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129777, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.937163] env[61991]: DEBUG nova.scheduler.client.report [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.202942] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129776, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.313428] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.445476] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.448167] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.632s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.449027] env[61991]: DEBUG nova.objects.instance [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lazy-loading 'resources' on Instance uuid 884e465b-e14f-4114-81a6-40e2e8694558 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.480466] env[61991]: INFO nova.scheduler.client.report [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted allocations for instance e3e5ec72-eec1-4155-bab2-038e3d5e38a8 [ 815.528287] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.557886] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.558187] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.558825] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.558825] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.558825] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.561058] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.562205] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.562205] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.563681] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.563766] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.563909] env[61991]: DEBUG nova.virt.hardware [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.565992] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2606738e-e834-4617-a386-38f31d77f92a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.578375] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a15f27-c944-4e73-99c1-960f8b8e32e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.703732] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129776, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.809918] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.990929] env[61991]: DEBUG oslo_concurrency.lockutils [None req-59b07923-448d-4f6b-bc0d-b4dba13a11b6 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "e3e5ec72-eec1-4155-bab2-038e3d5e38a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.572s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.139530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.139530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.206164] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129776, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.290478] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43a3808-5c5c-4230-a947-55e547f06b60 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.305125] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79036dc2-9b2e-4341-93a1-7b29d2054f7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.315609] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129777, 'name': ReconfigVM_Task, 'duration_secs': 1.213594} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.348335] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 49324fef-ad48-451b-a5ce-d9a1231137db/49324fef-ad48-451b-a5ce-d9a1231137db.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.349508] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-896e39a8-c0b8-4ba0-a00b-45a19a5cbbc6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.351812] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bf492a-0f6b-40d4-a5d3-20a7d68e1fec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.361702] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 816.361702] env[61991]: value = "task-1129778" [ 816.361702] env[61991]: _type = "Task" [ 816.361702] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.362968] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7521902c-d0f9-4dc0-9bb7-4b8d6f85f2cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.380018] env[61991]: DEBUG nova.compute.provider_tree [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.561397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "1715711a-72c5-4e86-88e0-cecf922cb42f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.561689] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.561893] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "1715711a-72c5-4e86-88e0-cecf922cb42f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.562236] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.562304] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.564508] env[61991]: INFO nova.compute.manager [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Terminating instance [ 816.566543] env[61991]: DEBUG nova.compute.manager [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.566750] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.567586] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b862f6-ba2c-49c3-85cf-7f7c5c35fde2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.575711] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.576115] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a4cb301-3ad7-42d4-b7f1-820154c6def6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.585338] env[61991]: DEBUG oslo_vmware.api [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 816.585338] env[61991]: value = "task-1129779" [ 816.585338] env[61991]: _type = "Task" [ 816.585338] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.595418] env[61991]: DEBUG oslo_vmware.api [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129779, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.642320] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.706443] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129776, 'name': CloneVM_Task, 'duration_secs': 1.733441} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.709612] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Created linked-clone VM from snapshot [ 816.709612] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357da073-bc00-4b42-a839-73408325be47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.720115] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Uploading image 340c522f-3717-4c6c-90f7-ab0ea3fc2355 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 816.755339] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 816.755339] env[61991]: value = "vm-246881" [ 816.755339] env[61991]: _type = "VirtualMachine" [ 816.755339] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 816.756530] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-51ecd3d9-ee89-425d-851d-654f35a0455d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.764141] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease: (returnval){ [ 816.764141] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5228658b-aa3a-f5e2-8cf4-ba5157b5a70a" [ 816.764141] env[61991]: _type = "HttpNfcLease" [ 816.764141] env[61991]: } obtained for exporting VM: (result){ [ 816.764141] env[61991]: value = "vm-246881" [ 816.764141] env[61991]: _type = "VirtualMachine" [ 816.764141] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 816.765030] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the lease: (returnval){ [ 816.765030] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5228658b-aa3a-f5e2-8cf4-ba5157b5a70a" [ 816.765030] env[61991]: _type = "HttpNfcLease" [ 816.765030] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 816.775832] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 816.775832] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5228658b-aa3a-f5e2-8cf4-ba5157b5a70a" [ 816.775832] env[61991]: _type = "HttpNfcLease" [ 816.775832] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 816.879203] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129778, 'name': Rename_Task, 'duration_secs': 0.147202} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.879617] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.879917] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a56c6631-21a5-4c11-af1e-71cb1db369ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.890907] env[61991]: DEBUG nova.scheduler.client.report [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.901572] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 816.901572] env[61991]: value = "task-1129781" [ 816.901572] env[61991]: _type = "Task" [ 816.901572] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.913289] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129781, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.024777] env[61991]: DEBUG nova.compute.manager [req-f6c23b51-eae4-4a7b-993a-e077f579b0c4 req-c673f409-b136-4e6d-ab56-5c3fcaafcaab service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Received event network-vif-plugged-d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.024854] env[61991]: DEBUG oslo_concurrency.lockutils [req-f6c23b51-eae4-4a7b-993a-e077f579b0c4 req-c673f409-b136-4e6d-ab56-5c3fcaafcaab service nova] Acquiring lock "15215bba-8160-408b-9e8f-afbb9e71d50e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.025254] env[61991]: DEBUG oslo_concurrency.lockutils [req-f6c23b51-eae4-4a7b-993a-e077f579b0c4 req-c673f409-b136-4e6d-ab56-5c3fcaafcaab service nova] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.025964] env[61991]: DEBUG oslo_concurrency.lockutils [req-f6c23b51-eae4-4a7b-993a-e077f579b0c4 req-c673f409-b136-4e6d-ab56-5c3fcaafcaab service nova] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.026290] env[61991]: DEBUG nova.compute.manager [req-f6c23b51-eae4-4a7b-993a-e077f579b0c4 req-c673f409-b136-4e6d-ab56-5c3fcaafcaab service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] No waiting events found dispatching network-vif-plugged-d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.026639] env[61991]: WARNING nova.compute.manager [req-f6c23b51-eae4-4a7b-993a-e077f579b0c4 req-c673f409-b136-4e6d-ab56-5c3fcaafcaab service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Received unexpected event network-vif-plugged-d450de43-1f93-4fc3-b519-0124646fd429 for instance with vm_state building and task_state spawning. [ 817.101166] env[61991]: DEBUG oslo_vmware.api [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129779, 'name': PowerOffVM_Task, 'duration_secs': 0.331773} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.101545] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 817.101765] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 817.102249] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c743f36-7d72-488f-be49-218fe524ffd9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.161885] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 817.162182] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 817.162316] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleting the datastore file [datastore2] 1715711a-72c5-4e86-88e0-cecf922cb42f {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.162647] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c5f7777-bf78-4b12-ac3f-876af0b8fbf7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.168213] env[61991]: DEBUG nova.network.neutron [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Successfully updated port: d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.171586] env[61991]: DEBUG oslo_vmware.api [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 817.171586] env[61991]: value = "task-1129783" [ 817.171586] env[61991]: _type = "Task" [ 817.171586] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.176964] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.182725] env[61991]: DEBUG oslo_vmware.api [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129783, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.273526] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 817.273526] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5228658b-aa3a-f5e2-8cf4-ba5157b5a70a" [ 817.273526] env[61991]: _type = "HttpNfcLease" [ 817.273526] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 817.273825] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 817.273825] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5228658b-aa3a-f5e2-8cf4-ba5157b5a70a" [ 817.273825] env[61991]: _type = "HttpNfcLease" [ 817.273825] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 817.274561] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded80e7d-1a1a-4c8c-9e09-79a93e806e2a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.282435] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eef4f3-3394-4fd9-a7ff-19bbc7e3fb62/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 817.282625] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eef4f3-3394-4fd9-a7ff-19bbc7e3fb62/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 817.388914] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.389074] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.403044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.409277] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-537f0b87-74e3-4d22-9101-9c01aceec278 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.411971] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.620s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.412446] env[61991]: DEBUG nova.objects.instance [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lazy-loading 'resources' on Instance uuid 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.428945] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129781, 'name': PowerOnVM_Task, 'duration_secs': 0.458927} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.430645] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.430842] env[61991]: DEBUG nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.434286] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee4a42f-62ea-4f2b-b214-7bfa7ff49b2b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.464209] env[61991]: INFO nova.scheduler.client.report [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Deleted allocations for instance 884e465b-e14f-4114-81a6-40e2e8694558 [ 817.673637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-15215bba-8160-408b-9e8f-afbb9e71d50e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.673637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-15215bba-8160-408b-9e8f-afbb9e71d50e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.673847] env[61991]: DEBUG nova.network.neutron [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 817.693332] env[61991]: DEBUG oslo_vmware.api [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129783, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264822} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.694031] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.695039] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.695039] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.695039] env[61991]: INFO nova.compute.manager [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 817.695529] env[61991]: DEBUG oslo.service.loopingcall [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.695763] env[61991]: DEBUG nova.compute.manager [-] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.695930] env[61991]: DEBUG nova.network.neutron [-] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 817.893379] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.955077] env[61991]: INFO nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] bringing vm to original state: 'stopped' [ 817.974416] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f78ffea-b69b-4523-ba8f-c21d9c5e2372 tempest-ServerMetadataTestJSON-1515976867 tempest-ServerMetadataTestJSON-1515976867-project-member] Lock "884e465b-e14f-4114-81a6-40e2e8694558" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.388s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.273567] env[61991]: DEBUG nova.network.neutron [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.296053] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83576752-0d42-49c2-a6b9-835cb2907535 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.304674] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b660493d-380e-4ce8-9c6f-3ae5d7f6a252 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.341960] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c07b798-f4fb-4541-b4dd-449ab8cb4fd6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.350689] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60703959-37a8-43b3-a77a-459c2bc661ae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.370232] env[61991]: DEBUG nova.compute.provider_tree [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.429655] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.874674] env[61991]: DEBUG nova.scheduler.client.report [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.951624] env[61991]: DEBUG nova.network.neutron [-] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.963078] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "49324fef-ad48-451b-a5ce-d9a1231137db" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.965970] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.003s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.966380] env[61991]: DEBUG nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.969027] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d33e78-4fed-461c-81a1-e8bd17ce6fce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.977906] env[61991]: DEBUG nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 818.981330] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.982252] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17b3c792-0020-4718-9b89-8cb150eb877a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.985993] env[61991]: DEBUG nova.network.neutron [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Updating instance_info_cache with network_info: [{"id": "d450de43-1f93-4fc3-b519-0124646fd429", "address": "fa:16:3e:02:f3:ee", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd450de43-1f", "ovs_interfaceid": "d450de43-1f93-4fc3-b519-0124646fd429", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.996864] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 818.996864] env[61991]: value = "task-1129784" [ 818.996864] env[61991]: _type = "Task" [ 818.996864] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.007481] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.158752] env[61991]: DEBUG nova.compute.manager [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Received event network-changed-d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.159390] env[61991]: DEBUG nova.compute.manager [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Refreshing instance network info cache due to event network-changed-d450de43-1f93-4fc3-b519-0124646fd429. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.159932] env[61991]: DEBUG oslo_concurrency.lockutils [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] Acquiring lock "refresh_cache-15215bba-8160-408b-9e8f-afbb9e71d50e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.385682] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.393044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.195s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.393927] env[61991]: DEBUG nova.objects.instance [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lazy-loading 'resources' on Instance uuid 671e0197-4f96-4660-be99-ea1d1c0588e1 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.432814] env[61991]: INFO nova.scheduler.client.report [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleted allocations for instance 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997 [ 819.456429] env[61991]: INFO nova.compute.manager [-] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Took 1.76 seconds to deallocate network for instance. [ 819.489374] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-15215bba-8160-408b-9e8f-afbb9e71d50e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.489720] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Instance network_info: |[{"id": "d450de43-1f93-4fc3-b519-0124646fd429", "address": "fa:16:3e:02:f3:ee", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd450de43-1f", "ovs_interfaceid": "d450de43-1f93-4fc3-b519-0124646fd429", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.490674] env[61991]: DEBUG oslo_concurrency.lockutils [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] Acquired lock "refresh_cache-15215bba-8160-408b-9e8f-afbb9e71d50e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.490868] env[61991]: DEBUG nova.network.neutron [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Refreshing network info cache for port d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.492637] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:f3:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd450de43-1f93-4fc3-b519-0124646fd429', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.500709] env[61991]: DEBUG oslo.service.loopingcall [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.505482] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 819.510021] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-945b008f-7ceb-48ca-a842-c8a10b2aae10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.533476] env[61991]: DEBUG oslo_vmware.api [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129784, 'name': PowerOffVM_Task, 'duration_secs': 0.291705} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.535133] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.535304] env[61991]: DEBUG nova.compute.manager [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.535495] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.535495] env[61991]: value = "task-1129785" [ 819.535495] env[61991]: _type = "Task" [ 819.535495] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.536243] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0b53a7-03c3-45a1-ad69-8cc85ba4f5af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.552131] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129785, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.951092] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4644a532-d1fb-438d-a7bc-fd29cf90a0bb tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "7cbf26bd-51ab-41f7-976a-4ad9c3c1a997" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.337s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.973459] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.057798] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129785, 'name': CreateVM_Task, 'duration_secs': 0.43409} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.057798] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.090s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.060446] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 820.061569] env[61991]: DEBUG nova.network.neutron [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Updated VIF entry in instance network info cache for port d450de43-1f93-4fc3-b519-0124646fd429. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.062450] env[61991]: DEBUG nova.network.neutron [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Updating instance_info_cache with network_info: [{"id": "d450de43-1f93-4fc3-b519-0124646fd429", "address": "fa:16:3e:02:f3:ee", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd450de43-1f", "ovs_interfaceid": "d450de43-1f93-4fc3-b519-0124646fd429", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.067146] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.067146] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.067146] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.067146] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61e252a2-0df2-4d72-9102-98761efd20d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.072945] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 820.072945] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e59fe0-e3a6-1fe3-354c-dff83744ca52" [ 820.072945] env[61991]: _type = "Task" [ 820.072945] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.084205] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e59fe0-e3a6-1fe3-354c-dff83744ca52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.234350] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490373ca-172d-4ff4-9b4a-4a41c9b1d847 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.242733] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a81f1ef-47df-4c6e-b76a-391a915f51b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.275847] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951a726f-7820-42b4-b47f-8984202a5553 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.284920] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295afac0-3711-4be1-8873-513a9acab312 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.299055] env[61991]: DEBUG nova.compute.provider_tree [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.569468] env[61991]: DEBUG oslo_concurrency.lockutils [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] Releasing lock "refresh_cache-15215bba-8160-408b-9e8f-afbb9e71d50e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.569790] env[61991]: DEBUG nova.compute.manager [req-98f6f2cb-9093-45c5-85d2-fc91fe7fa67d req-bf63b190-b2e4-4f6a-9a28-9c240999ba72 service nova] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Received event network-vif-deleted-91eaaf4e-f26e-42c4-a3da-3a34c4c22e4f {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.578347] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.585566] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e59fe0-e3a6-1fe3-354c-dff83744ca52, 'name': SearchDatastore_Task, 'duration_secs': 0.013888} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.585986] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.586355] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.587199] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.587917] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.588232] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.588539] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b7e3f16-84ed-4670-9673-466454d00ecb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.599155] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.599417] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.600718] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dec47881-e25c-4b37-ae2e-4ddf95484946 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.608835] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 820.608835] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eaf396-aa5b-6c74-e7ee-283a1ba203f3" [ 820.608835] env[61991]: _type = "Task" [ 820.608835] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.619021] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eaf396-aa5b-6c74-e7ee-283a1ba203f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.803760] env[61991]: DEBUG nova.scheduler.client.report [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.836222] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "3183190e-3202-40be-894c-e47b186708b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.836222] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "3183190e-3202-40be-894c-e47b186708b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.123440] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eaf396-aa5b-6c74-e7ee-283a1ba203f3, 'name': SearchDatastore_Task, 'duration_secs': 0.011683} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.125297] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce1d4413-c1de-4b10-99d9-9a3c443fe999 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.131319] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 821.131319] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523d47fe-c12c-4a59-bb3a-d83c18058248" [ 821.131319] env[61991]: _type = "Task" [ 821.131319] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.140602] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523d47fe-c12c-4a59-bb3a-d83c18058248, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.310555] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.917s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.313123] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.487s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.313412] env[61991]: DEBUG nova.objects.instance [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lazy-loading 'pci_requests' on Instance uuid f78ef63b-453e-45d3-959b-4b0c1922b53e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.343158] env[61991]: INFO nova.scheduler.client.report [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Deleted allocations for instance 671e0197-4f96-4660-be99-ea1d1c0588e1 [ 821.347111] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.602309] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "49324fef-ad48-451b-a5ce-d9a1231137db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.602777] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.604575] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "49324fef-ad48-451b-a5ce-d9a1231137db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.604861] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.605086] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.609497] env[61991]: INFO nova.compute.manager [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Terminating instance [ 821.611641] env[61991]: DEBUG nova.compute.manager [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.611838] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.612699] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2ed9cf-43be-41ee-ab9b-7d9397683b7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.621431] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 821.621701] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2fe8a980-1d41-4dd4-8ff2-4c41842e8e16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.641449] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523d47fe-c12c-4a59-bb3a-d83c18058248, 'name': SearchDatastore_Task, 'duration_secs': 0.013657} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.641588] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.644101] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 15215bba-8160-408b-9e8f-afbb9e71d50e/15215bba-8160-408b-9e8f-afbb9e71d50e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.644101] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dacec6ee-4a9e-45d1-8a7f-43491b5cbc41 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.648745] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 821.648745] env[61991]: value = "task-1129787" [ 821.648745] env[61991]: _type = "Task" [ 821.648745] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.658050] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129787, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.687758] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 821.690025] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 821.690025] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore1] 49324fef-ad48-451b-a5ce-d9a1231137db {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.690025] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7099803-7554-44bd-84ae-238eb3ab04a4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.695907] env[61991]: DEBUG oslo_vmware.api [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 821.695907] env[61991]: value = "task-1129788" [ 821.695907] env[61991]: _type = "Task" [ 821.695907] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.705271] env[61991]: DEBUG oslo_vmware.api [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129788, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.821403] env[61991]: DEBUG nova.objects.instance [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lazy-loading 'numa_topology' on Instance uuid f78ef63b-453e-45d3-959b-4b0c1922b53e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.859577] env[61991]: DEBUG oslo_concurrency.lockutils [None req-153b934d-358e-40a6-8a77-0a8d478d09c1 tempest-ListServersNegativeTestJSON-548535671 tempest-ListServersNegativeTestJSON-548535671-project-member] Lock "671e0197-4f96-4660-be99-ea1d1c0588e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.038s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.884682] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.162193] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129787, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.207787] env[61991]: DEBUG oslo_vmware.api [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129788, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296812} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.208137] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 822.208356] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 822.208567] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.208776] env[61991]: INFO nova.compute.manager [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Took 0.60 seconds to destroy the instance on the hypervisor. [ 822.209092] env[61991]: DEBUG oslo.service.loopingcall [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.209318] env[61991]: DEBUG nova.compute.manager [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.209427] env[61991]: DEBUG nova.network.neutron [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 822.327314] env[61991]: INFO nova.compute.claims [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.496987] env[61991]: DEBUG nova.compute.manager [req-c790803f-781d-493c-b7eb-192b276a6ad7 req-ff1af503-0273-4836-b897-c85da268b085 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Received event network-vif-deleted-fa5c2434-a034-4ee0-b847-7108e636ef8d {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.498410] env[61991]: INFO nova.compute.manager [req-c790803f-781d-493c-b7eb-192b276a6ad7 req-ff1af503-0273-4836-b897-c85da268b085 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Neutron deleted interface fa5c2434-a034-4ee0-b847-7108e636ef8d; detaching it from the instance and deleting it from the info cache [ 822.498763] env[61991]: DEBUG nova.network.neutron [req-c790803f-781d-493c-b7eb-192b276a6ad7 req-ff1af503-0273-4836-b897-c85da268b085 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.660761] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129787, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575926} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.661071] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 15215bba-8160-408b-9e8f-afbb9e71d50e/15215bba-8160-408b-9e8f-afbb9e71d50e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.661302] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.661569] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd96a761-684f-4e10-a9db-950dad708a29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.668644] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 822.668644] env[61991]: value = "task-1129789" [ 822.668644] env[61991]: _type = "Task" [ 822.668644] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.683654] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.972524] env[61991]: DEBUG nova.network.neutron [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.001726] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6cd0436-858a-4949-83e7-3468d2e07f14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.014654] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17a2d77-06ae-459a-805b-c6276a077d87 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.043315] env[61991]: DEBUG nova.compute.manager [req-c790803f-781d-493c-b7eb-192b276a6ad7 req-ff1af503-0273-4836-b897-c85da268b085 service nova] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Detach interface failed, port_id=fa5c2434-a034-4ee0-b847-7108e636ef8d, reason: Instance 49324fef-ad48-451b-a5ce-d9a1231137db could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 823.178785] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078894} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.179075] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.179821] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49aa34a1-996b-42e8-9232-88ed11535eeb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.208355] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 15215bba-8160-408b-9e8f-afbb9e71d50e/15215bba-8160-408b-9e8f-afbb9e71d50e.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.208710] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-376c31d1-d4ca-4538-9830-88bc2b0e3dd8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.230944] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 823.230944] env[61991]: value = "task-1129790" [ 823.230944] env[61991]: _type = "Task" [ 823.230944] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.239483] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129790, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.475611] env[61991]: INFO nova.compute.manager [-] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Took 1.27 seconds to deallocate network for instance. [ 823.567984] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94a13ed-082f-42c4-af26-6bee84d016d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.576062] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62070fe4-0de2-410c-9567-bb86bbaf7313 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.609917] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ff1f39-91b2-459c-bd78-5f56f60dc2f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.618367] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f459a5f8-2331-4a91-95a8-257b0141ac89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.632214] env[61991]: DEBUG nova.compute.provider_tree [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.744076] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129790, 'name': ReconfigVM_Task, 'duration_secs': 0.427058} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.744076] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 15215bba-8160-408b-9e8f-afbb9e71d50e/15215bba-8160-408b-9e8f-afbb9e71d50e.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.744076] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea5fb7bb-7716-4028-9786-d0d013535128 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.750072] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 823.750072] env[61991]: value = "task-1129791" [ 823.750072] env[61991]: _type = "Task" [ 823.750072] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.759927] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129791, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.982607] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.135366] env[61991]: DEBUG nova.scheduler.client.report [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.262319] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129791, 'name': Rename_Task, 'duration_secs': 0.186261} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.262600] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.262866] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6a9e054-224d-4736-9faa-37c0424bb98a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.270019] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 824.270019] env[61991]: value = "task-1129792" [ 824.270019] env[61991]: _type = "Task" [ 824.270019] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.278678] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.641705] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.329s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.646185] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.085s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.646185] env[61991]: DEBUG nova.objects.instance [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'resources' on Instance uuid 80d05278-b516-4408-94b0-11bc93500b5c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.728778] env[61991]: INFO nova.network.neutron [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating port 4b255035-ebfa-4e8a-b9e5-a51737f080b9 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 824.786428] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129792, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.284916] env[61991]: DEBUG oslo_vmware.api [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129792, 'name': PowerOnVM_Task, 'duration_secs': 0.576002} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.285222] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.285575] env[61991]: INFO nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Took 9.76 seconds to spawn the instance on the hypervisor. [ 825.285649] env[61991]: DEBUG nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.289519] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce5b0a2-523d-4b67-9bd2-fe82b6e2510e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.404315] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5248f31a-581b-41af-b4e1-3b31a678ec80 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.412271] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce15674-6828-4432-9e96-5c5eaa272bc9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.444655] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b819793-b349-408a-a60d-83b2b4a3c101 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.452442] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4f191c-d968-426b-b7a4-0eb93cb04516 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.467446] env[61991]: DEBUG nova.compute.provider_tree [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.815290] env[61991]: INFO nova.compute.manager [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Took 32.51 seconds to build instance. [ 825.971161] env[61991]: DEBUG nova.scheduler.client.report [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.317356] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3841904a-c118-4608-bc68-6e997a77801b tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.017s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.475881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.478382] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.224s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.478615] env[61991]: DEBUG nova.objects.instance [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lazy-loading 'resources' on Instance uuid 3e62f219-7fec-470f-9d0a-e292d62e076c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.507989] env[61991]: INFO nova.scheduler.client.report [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted allocations for instance 80d05278-b516-4408-94b0-11bc93500b5c [ 826.532793] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eef4f3-3394-4fd9-a7ff-19bbc7e3fb62/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 826.533773] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb32f66-eea3-4a26-bc1c-41da72dfa0e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.546259] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eef4f3-3394-4fd9-a7ff-19bbc7e3fb62/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 826.546259] env[61991]: ERROR oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eef4f3-3394-4fd9-a7ff-19bbc7e3fb62/disk-0.vmdk due to incomplete transfer. [ 826.546259] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-677a6d51-62e9-4e73-839b-0411980d05f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.552043] env[61991]: DEBUG oslo_vmware.rw_handles [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eef4f3-3394-4fd9-a7ff-19bbc7e3fb62/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 826.552372] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Uploaded image 340c522f-3717-4c6c-90f7-ab0ea3fc2355 to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 826.556697] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 826.556812] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0533574c-68fe-4897-959b-1c6495e7150f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.566896] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 826.566896] env[61991]: value = "task-1129793" [ 826.566896] env[61991]: _type = "Task" [ 826.566896] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.576067] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129793, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.612945] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "15215bba-8160-408b-9e8f-afbb9e71d50e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.613238] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.613427] env[61991]: DEBUG nova.compute.manager [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.614642] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0024214-3f90-4fb7-9c49-553855ebf6ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.623825] env[61991]: DEBUG nova.compute.manager [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 826.623825] env[61991]: DEBUG nova.objects.instance [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lazy-loading 'flavor' on Instance uuid 15215bba-8160-408b-9e8f-afbb9e71d50e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.863438] env[61991]: DEBUG nova.compute.manager [req-029bc5d6-96c5-4aa6-a617-0f92a0ee4d78 req-125039b1-0b5b-490c-a72e-9b252e468f38 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-vif-plugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.863847] env[61991]: DEBUG oslo_concurrency.lockutils [req-029bc5d6-96c5-4aa6-a617-0f92a0ee4d78 req-125039b1-0b5b-490c-a72e-9b252e468f38 service nova] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.864180] env[61991]: DEBUG oslo_concurrency.lockutils [req-029bc5d6-96c5-4aa6-a617-0f92a0ee4d78 req-125039b1-0b5b-490c-a72e-9b252e468f38 service nova] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.864396] env[61991]: DEBUG oslo_concurrency.lockutils [req-029bc5d6-96c5-4aa6-a617-0f92a0ee4d78 req-125039b1-0b5b-490c-a72e-9b252e468f38 service nova] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.864720] env[61991]: DEBUG nova.compute.manager [req-029bc5d6-96c5-4aa6-a617-0f92a0ee4d78 req-125039b1-0b5b-490c-a72e-9b252e468f38 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] No waiting events found dispatching network-vif-plugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 826.864989] env[61991]: WARNING nova.compute.manager [req-029bc5d6-96c5-4aa6-a617-0f92a0ee4d78 req-125039b1-0b5b-490c-a72e-9b252e468f38 service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received unexpected event network-vif-plugged-4b255035-ebfa-4e8a-b9e5-a51737f080b9 for instance with vm_state shelved_offloaded and task_state spawning. [ 826.898609] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.898609] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.898609] env[61991]: DEBUG nova.network.neutron [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 827.020756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b1f391ea-a9e6-4215-9304-65c257731eb0 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "80d05278-b516-4408-94b0-11bc93500b5c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.254s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.021712] env[61991]: DEBUG oslo_concurrency.lockutils [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] Acquired lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.022913] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eddf04-d2c4-4d58-b59c-5ddf406278a4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.034405] env[61991]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 827.034650] env[61991]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61991) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 827.035135] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8eb43a31-bb28-4b89-846d-24eab0130b6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.047275] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d219d524-4541-4a9b-a549-4c9cce8dd378 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.084098] env[61991]: ERROR root [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-246827' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-246827' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-246827' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-246827'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-246827' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-246827' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-246827'}\n"]: nova.exception.InstanceNotFound: Instance 80d05278-b516-4408-94b0-11bc93500b5c could not be found. [ 827.087632] env[61991]: DEBUG oslo_concurrency.lockutils [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] Releasing lock "80d05278-b516-4408-94b0-11bc93500b5c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.087632] env[61991]: DEBUG nova.compute.manager [req-437383b4-e346-4a51-868f-46f78c39a9c0 req-71c27011-dbda-4cf5-b030-c45ad5859ac8 service nova] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Detach interface failed, port_id=e08c38be-49fa-4a29-9cb1-37e15930e258, reason: Instance 80d05278-b516-4408-94b0-11bc93500b5c could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 827.095109] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129793, 'name': Destroy_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.129323] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 827.129650] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c905d772-f607-417f-a715-18cc97565dae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.138463] env[61991]: DEBUG oslo_vmware.api [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 827.138463] env[61991]: value = "task-1129794" [ 827.138463] env[61991]: _type = "Task" [ 827.138463] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.153799] env[61991]: DEBUG oslo_vmware.api [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129794, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.286490] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c3997e-23c9-4bf4-a046-aeb77088f3c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.294507] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1d7e1b-740f-404f-91a6-c7e62e5a1022 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.326327] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a415b4-6a6c-415d-9f8f-ca4c6214b16f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.334050] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5911571-cd15-407e-8a3c-96ba8ba14032 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.349898] env[61991]: DEBUG nova.compute.provider_tree [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.577632] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129793, 'name': Destroy_Task, 'duration_secs': 0.707519} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.577961] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroyed the VM [ 827.578233] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 827.578505] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d462a9dd-b03f-4e50-b057-3130b4e7cac0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.588211] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 827.588211] env[61991]: value = "task-1129795" [ 827.588211] env[61991]: _type = "Task" [ 827.588211] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.599138] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129795, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.650773] env[61991]: DEBUG oslo_vmware.api [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129794, 'name': PowerOffVM_Task, 'duration_secs': 0.211947} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.651070] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.651240] env[61991]: DEBUG nova.compute.manager [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.652140] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b563b61-0df3-4c51-a5c5-7b673029b683 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.665924] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.666115] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.780742] env[61991]: DEBUG nova.network.neutron [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.851940] env[61991]: DEBUG nova.scheduler.client.report [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.098905] env[61991]: DEBUG oslo_vmware.api [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129795, 'name': RemoveSnapshot_Task, 'duration_secs': 0.466138} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.098905] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 828.099061] env[61991]: INFO nova.compute.manager [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 15.52 seconds to snapshot the instance on the hypervisor. [ 828.159759] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "f7eab1a9-2a8a-4664-8736-ca599e207c31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.159993] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.169174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f989adc7-98ad-44b4-91a0-03d77e7eb22f tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.170693] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.284490] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.315041] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6b503e1e8cdc6260e56ef7a91c19f6bf',container_format='bare',created_at=2024-10-04T09:21:34Z,direct_url=,disk_format='vmdk',id=0810354d-ca0b-4776-8487-42797b713043,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-288634761-shelved',owner='d7b25d4745e1419b963e387c8461dd7d',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-04T09:21:51Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.315822] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.316138] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.317022] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.317022] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.317022] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.317022] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.317222] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.317296] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.317465] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.317640] env[61991]: DEBUG nova.virt.hardware [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.318739] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c268ddcd-21ee-4497-a699-b451161e6e23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.327197] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a049d886-7cb0-44a9-bf50-e165a40de394 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.343452] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:89:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a94c29-ddd5-4383-9219-1c2c3bb09cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b255035-ebfa-4e8a-b9e5-a51737f080b9', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.351297] env[61991]: DEBUG oslo.service.loopingcall [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.351958] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.352233] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ef9602d-05f8-4579-9d3e-9910dfcf2e8b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.368086] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.370554] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.244s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.372091] env[61991]: INFO nova.compute.claims [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.380267] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.380267] env[61991]: value = "task-1129796" [ 828.380267] env[61991]: _type = "Task" [ 828.380267] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.393218] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129796, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.399160] env[61991]: INFO nova.scheduler.client.report [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleted allocations for instance 3e62f219-7fec-470f-9d0a-e292d62e076c [ 828.652481] env[61991]: DEBUG nova.compute.manager [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Found 3 images (rotation: 2) {{(pid=61991) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 828.652690] env[61991]: DEBUG nova.compute.manager [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Rotating out 1 backups {{(pid=61991) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 828.652921] env[61991]: DEBUG nova.compute.manager [None req-2318fda3-0c80-4054-b2b1-31bc0449e777 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleting image 5b317551-bad7-4685-b760-e9c291dff08c {{(pid=61991) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 828.663835] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.695559] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.895018] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129796, 'name': CreateVM_Task, 'duration_secs': 0.336957} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.895018] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.895018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.895018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.895018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.895018] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf4b336d-a63c-48e5-b81d-967422dec1e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.898569] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 828.898569] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e70b82-21d6-f57f-714f-6292cb41d0d2" [ 828.898569] env[61991]: _type = "Task" [ 828.898569] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.908566] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e70b82-21d6-f57f-714f-6292cb41d0d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.909039] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8281379a-dc07-4a05-abcf-2afcbe1b7091 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "3e62f219-7fec-470f-9d0a-e292d62e076c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.449s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.118345] env[61991]: DEBUG nova.compute.manager [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.118345] env[61991]: DEBUG nova.compute.manager [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing instance network info cache due to event network-changed-4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.118345] env[61991]: DEBUG oslo_concurrency.lockutils [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] Acquiring lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.118345] env[61991]: DEBUG oslo_concurrency.lockutils [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] Acquired lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.118345] env[61991]: DEBUG nova.network.neutron [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Refreshing network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.164436] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "15215bba-8160-408b-9e8f-afbb9e71d50e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.164709] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.164967] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "15215bba-8160-408b-9e8f-afbb9e71d50e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.165223] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.165362] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.172103] env[61991]: INFO nova.compute.manager [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Terminating instance [ 829.175872] env[61991]: DEBUG nova.compute.manager [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.176113] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.176915] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293ff32b-7871-4cd6-8e4f-7e60e7117887 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.188041] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.188306] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-561637fb-d04b-4c0e-b94e-21a64431cdcb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.200522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.265118] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.265396] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.265591] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleting the datastore file [datastore1] 15215bba-8160-408b-9e8f-afbb9e71d50e {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.265862] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb1c2aaa-96c0-4651-bf58-036a013edee3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.274628] env[61991]: DEBUG oslo_vmware.api [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 829.274628] env[61991]: value = "task-1129798" [ 829.274628] env[61991]: _type = "Task" [ 829.274628] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.283035] env[61991]: DEBUG oslo_vmware.api [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129798, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.412919] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.413551] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Processing image 0810354d-ca0b-4776-8487-42797b713043 {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.413551] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.413680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.413782] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.414066] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6651ce5b-2b19-4018-a39b-e7cd3373ead3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.431917] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.432265] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.435476] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c878fc5-4b73-4545-8278-377b9554d0dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.441609] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 829.441609] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527370eb-7b5b-d5e4-382d-984a0a89ec13" [ 829.441609] env[61991]: _type = "Task" [ 829.441609] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.449237] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527370eb-7b5b-d5e4-382d-984a0a89ec13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.655400] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9b42ee-3a05-4cc3-b85d-5f2192b1b883 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.663947] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f0810f-dcb7-45b8-a64b-86349af92d31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.697060] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fadac13-799c-4deb-b99b-be7b0cdc9441 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.704822] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdb971a-fe62-4be0-8c41-c4539ca40032 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.719518] env[61991]: DEBUG nova.compute.provider_tree [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.724174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.724387] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.785605] env[61991]: DEBUG oslo_vmware.api [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129798, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151047} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.785884] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.786106] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.786285] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.786485] env[61991]: INFO nova.compute.manager [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Took 0.61 seconds to destroy the instance on the hypervisor. [ 829.786717] env[61991]: DEBUG oslo.service.loopingcall [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.786898] env[61991]: DEBUG nova.compute.manager [-] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.786985] env[61991]: DEBUG nova.network.neutron [-] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.622290] env[61991]: DEBUG nova.scheduler.client.report [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.625947] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.629240] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.629486] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.639046] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.639196] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.640496] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "97913aea-48c1-4fda-aee3-578e89a08f71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.640933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "97913aea-48c1-4fda-aee3-578e89a08f71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.651018] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Preparing fetch location {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 830.651018] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Fetch image to [datastore1] OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68/OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68.vmdk {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 830.651018] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Downloading stream optimized image 0810354d-ca0b-4776-8487-42797b713043 to [datastore1] OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68/OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68.vmdk on the data store datastore1 as vApp {{(pid=61991) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 830.651018] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Downloading image file data 0810354d-ca0b-4776-8487-42797b713043 to the ESX as VM named 'OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68' {{(pid=61991) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 830.730035] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 830.730035] env[61991]: value = "resgroup-9" [ 830.730035] env[61991]: _type = "ResourcePool" [ 830.730035] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 830.730376] env[61991]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7f955a96-6e8b-4130-8d1f-ab212b979d58 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.753262] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lease: (returnval){ [ 830.753262] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529982b1-a0e0-8d4f-fffc-f8b8c96d63f5" [ 830.753262] env[61991]: _type = "HttpNfcLease" [ 830.753262] env[61991]: } obtained for vApp import into resource pool (val){ [ 830.753262] env[61991]: value = "resgroup-9" [ 830.753262] env[61991]: _type = "ResourcePool" [ 830.753262] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 830.753670] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the lease: (returnval){ [ 830.753670] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529982b1-a0e0-8d4f-fffc-f8b8c96d63f5" [ 830.753670] env[61991]: _type = "HttpNfcLease" [ 830.753670] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 830.764966] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 830.764966] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529982b1-a0e0-8d4f-fffc-f8b8c96d63f5" [ 830.764966] env[61991]: _type = "HttpNfcLease" [ 830.764966] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 830.798062] env[61991]: DEBUG nova.network.neutron [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updated VIF entry in instance network info cache for port 4b255035-ebfa-4e8a-b9e5-a51737f080b9. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.798465] env[61991]: DEBUG nova.network.neutron [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [{"id": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "address": "fa:16:3e:96:89:be", "network": {"id": "919af812-2234-41f2-9bbf-a2f75d8fdb76", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-372925114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7b25d4745e1419b963e387c8461dd7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a94c29-ddd5-4383-9219-1c2c3bb09cc5", "external-id": "nsx-vlan-transportzone-2", "segmentation_id": 2, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b255035-eb", "ovs_interfaceid": "4b255035-ebfa-4e8a-b9e5-a51737f080b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.122597] env[61991]: DEBUG nova.network.neutron [-] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.127077] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.756s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.127766] env[61991]: DEBUG nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.131616] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.939s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.133056] env[61991]: INFO nova.compute.claims [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.135838] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.140688] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 831.155285] env[61991]: DEBUG nova.compute.manager [req-22781a67-4102-44bd-af25-5081aa503288 req-5438c6aa-74ee-4e63-a5c8-62f8f3c2e649 service nova] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Received event network-vif-deleted-d450de43-1f93-4fc3-b519-0124646fd429 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.162936] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.262792] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 831.262792] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529982b1-a0e0-8d4f-fffc-f8b8c96d63f5" [ 831.262792] env[61991]: _type = "HttpNfcLease" [ 831.262792] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 831.263100] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 831.263100] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529982b1-a0e0-8d4f-fffc-f8b8c96d63f5" [ 831.263100] env[61991]: _type = "HttpNfcLease" [ 831.263100] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 831.263803] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1f0ba5-58b8-4b86-828e-8c7ab631ca12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.272079] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc8c5a-4a15-a254-92dc-8eb8b1460c59/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 831.272270] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc8c5a-4a15-a254-92dc-8eb8b1460c59/disk-0.vmdk. {{(pid=61991) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 831.330257] env[61991]: DEBUG oslo_concurrency.lockutils [req-7020ac73-c41f-4a53-94c9-bcbba5164d11 req-76f2e20b-f013-42cb-9ad8-5c0bf92ada0b service nova] Releasing lock "refresh_cache-f78ef63b-453e-45d3-959b-4b0c1922b53e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.336439] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-624f9c7b-31b7-40df-86bd-821a703ad774 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.537365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.537639] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.537825] env[61991]: DEBUG nova.compute.manager [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.538781] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e71336-ab2a-4bbc-8a53-c43b84702963 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.547738] env[61991]: DEBUG nova.compute.manager [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 831.547887] env[61991]: DEBUG nova.objects.instance [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'flavor' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.625223] env[61991]: INFO nova.compute.manager [-] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Took 1.84 seconds to deallocate network for instance. [ 831.638302] env[61991]: DEBUG nova.compute.utils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.640074] env[61991]: DEBUG nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 831.666648] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.667747] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.056045] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.056400] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6184e3a-2024-4c7f-b210-e239dbe68b14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.063555] env[61991]: DEBUG oslo_vmware.api [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 832.063555] env[61991]: value = "task-1129800" [ 832.063555] env[61991]: _type = "Task" [ 832.063555] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.072176] env[61991]: DEBUG oslo_vmware.api [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.133071] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.141092] env[61991]: DEBUG nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.453273] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94903f53-b385-43c5-857b-13c268dd4e96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.460646] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04103ed4-e70e-453d-b2be-995f06eade1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.491756] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521afb71-d98e-44c4-a7e4-2ab57164364e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.499265] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b366ca-e41a-4a63-9af0-5d4b59040d04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.513742] env[61991]: DEBUG nova.compute.provider_tree [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.574750] env[61991]: DEBUG oslo_vmware.api [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129800, 'name': PowerOffVM_Task, 'duration_secs': 0.227813} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.575173] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.575297] env[61991]: DEBUG nova.compute.manager [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.576129] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565da0b4-a153-4e42-88aa-082c4123f79b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.586708] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Completed reading data from the image iterator. {{(pid=61991) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 832.587437] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc8c5a-4a15-a254-92dc-8eb8b1460c59/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 832.587845] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7303b5dc-5b82-4ecb-88e2-a22d933aeaae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.595769] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc8c5a-4a15-a254-92dc-8eb8b1460c59/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 832.595769] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc8c5a-4a15-a254-92dc-8eb8b1460c59/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 832.595920] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-fd6cd0cc-2430-487a-9418-c699d03351fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.793367] env[61991]: DEBUG oslo_vmware.rw_handles [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc8c5a-4a15-a254-92dc-8eb8b1460c59/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 832.793648] env[61991]: INFO nova.virt.vmwareapi.images [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Downloaded image file data 0810354d-ca0b-4776-8487-42797b713043 [ 832.794528] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b189a5-ad0e-4f5f-845b-d58def9658d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.812730] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f084759d-2101-4751-84cb-3c3065a35006 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.849042] env[61991]: INFO nova.virt.vmwareapi.images [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] The imported VM was unregistered [ 832.851423] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Caching image {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 832.851717] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Creating directory with path [datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043 {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.852028] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c517fb76-f94a-4ead-b61c-4097826d2a44 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.865586] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Created directory with path [datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043 {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.865714] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68/OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68.vmdk to [datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk. {{(pid=61991) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 832.866013] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-75c54ac4-2498-46b9-bbaf-ff5f062c4e7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.872091] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 832.872091] env[61991]: value = "task-1129802" [ 832.872091] env[61991]: _type = "Task" [ 832.872091] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.879819] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129802, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.016653] env[61991]: DEBUG nova.scheduler.client.report [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.094643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2cea7bf1-e026-40e9-b282-8c753d9bd5e5 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.557s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.158494] env[61991]: DEBUG nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.187348] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.187815] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.188125] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.188409] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.188666] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.188925] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.189276] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.189560] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.189792] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.189999] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.190339] env[61991]: DEBUG nova.virt.hardware [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.191426] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca665adb-8008-481f-941b-c11fb40fbc8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.200727] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f70a497-f704-4acd-aafb-6ad576fe0ac3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.217395] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.224137] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Creating folder: Project (1b5c474cf16e4693b97fc4a961ae6027). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.224587] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-72dd39d0-c0f2-4c55-967d-b2d99d50bed9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.237012] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Created folder: Project (1b5c474cf16e4693b97fc4a961ae6027) in parent group-v246753. [ 833.237245] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Creating folder: Instances. Parent ref: group-v246885. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.237509] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a12c85e0-4ca6-49bd-9232-aaebe2a29802 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.248602] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Created folder: Instances in parent group-v246885. [ 833.248964] env[61991]: DEBUG oslo.service.loopingcall [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.249181] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.249413] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3e5e0e3-36c0-4f4f-870e-3b462777172b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.270374] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.270374] env[61991]: value = "task-1129805" [ 833.270374] env[61991]: _type = "Task" [ 833.270374] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.280348] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129805, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.385264] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129802, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.521243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.521864] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.528029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.348s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.528029] env[61991]: INFO nova.compute.claims [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.782265] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129805, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.885120] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129802, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.033958] env[61991]: DEBUG nova.compute.utils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.039030] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.039030] env[61991]: DEBUG nova.network.neutron [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 834.081668] env[61991]: DEBUG nova.policy [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af6cfe94f59140f1b85bf2a1e8593e70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79c013b6c254403b91f520e2af18027f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.285343] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129805, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.388363] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129802, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.396033] env[61991]: DEBUG nova.network.neutron [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Successfully created port: 87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.538898] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.797660] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129805, 'name': CreateVM_Task, 'duration_secs': 1.32377} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.797863] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.798355] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.798530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.798874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.799174] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c618152a-22d2-4f91-a29f-3c1a729e9afb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.814971] env[61991]: DEBUG nova.compute.manager [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Stashing vm_state: stopped {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 834.819975] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 834.819975] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eb4e0d-f01a-f316-e106-1d196ea441b4" [ 834.819975] env[61991]: _type = "Task" [ 834.819975] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.835482] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eb4e0d-f01a-f316-e106-1d196ea441b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.890717] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129802, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.928480] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94206e4f-6d9e-4d23-ad27-ef948378ea5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.938381] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05f30e5-3474-4edc-98f6-e5d61744daba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.970718] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8731d7ca-1194-4b2e-b863-226fee92a63c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.981294] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a0f18b-33cb-44ac-85e2-21c5058762e7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.997636] env[61991]: DEBUG nova.compute.provider_tree [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.334268] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52eb4e0d-f01a-f316-e106-1d196ea441b4, 'name': SearchDatastore_Task, 'duration_secs': 0.081478} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.335325] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.335618] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.335847] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.336150] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.336336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.336528] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.336780] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0af46cbc-1d99-4d7c-b951-c78003b3be4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.345245] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.345417] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.346114] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11087d16-2851-41b6-84e8-4f12f23ccf84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.350814] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 835.350814] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea3c57-82ed-e901-8be3-baf828ef49ef" [ 835.350814] env[61991]: _type = "Task" [ 835.350814] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.357966] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea3c57-82ed-e901-8be3-baf828ef49ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.386030] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129802, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.387503} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.386528] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68/OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68.vmdk to [datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk. [ 835.386528] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Cleaning up location [datastore1] OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68 {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 835.386635] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_eb6371b2-01ee-4fb5-a182-0c2f97382f68 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.386825] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b702c85-9efa-454c-8b9c-2eefe89945dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.392170] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 835.392170] env[61991]: value = "task-1129806" [ 835.392170] env[61991]: _type = "Task" [ 835.392170] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.399261] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.500960] env[61991]: DEBUG nova.scheduler.client.report [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.554526] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.580892] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.581215] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.582032] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.582032] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.582032] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.582032] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.582032] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.582243] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.582364] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.582467] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.582634] env[61991]: DEBUG nova.virt.hardware [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.583520] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d521344e-ec25-4306-ad55-704b4bc256b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.591410] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89c2174-bc69-47e3-9a80-4db2d8a8f730 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.798716] env[61991]: DEBUG nova.compute.manager [req-67dd4a74-26ef-491d-80b3-59f2a0d74109 req-f1e2f6c1-7cbc-4af4-8f2a-b4869574de93 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Received event network-vif-plugged-87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.798896] env[61991]: DEBUG oslo_concurrency.lockutils [req-67dd4a74-26ef-491d-80b3-59f2a0d74109 req-f1e2f6c1-7cbc-4af4-8f2a-b4869574de93 service nova] Acquiring lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.799081] env[61991]: DEBUG oslo_concurrency.lockutils [req-67dd4a74-26ef-491d-80b3-59f2a0d74109 req-f1e2f6c1-7cbc-4af4-8f2a-b4869574de93 service nova] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.799250] env[61991]: DEBUG oslo_concurrency.lockutils [req-67dd4a74-26ef-491d-80b3-59f2a0d74109 req-f1e2f6c1-7cbc-4af4-8f2a-b4869574de93 service nova] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.799442] env[61991]: DEBUG nova.compute.manager [req-67dd4a74-26ef-491d-80b3-59f2a0d74109 req-f1e2f6c1-7cbc-4af4-8f2a-b4869574de93 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] No waiting events found dispatching network-vif-plugged-87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.799585] env[61991]: WARNING nova.compute.manager [req-67dd4a74-26ef-491d-80b3-59f2a0d74109 req-f1e2f6c1-7cbc-4af4-8f2a-b4869574de93 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Received unexpected event network-vif-plugged-87e2f32e-53df-49a9-bae3-f9529f2d3c6c for instance with vm_state building and task_state spawning. [ 835.860770] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ea3c57-82ed-e901-8be3-baf828ef49ef, 'name': SearchDatastore_Task, 'duration_secs': 0.009739} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.861550] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8141b2f5-1d28-43a8-92e5-330340592a04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.866916] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 835.866916] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52099bf4-6497-4b8b-9ef6-c0a4b55da63a" [ 835.866916] env[61991]: _type = "Task" [ 835.866916] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.874664] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52099bf4-6497-4b8b-9ef6-c0a4b55da63a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.882795] env[61991]: DEBUG nova.network.neutron [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Successfully updated port: 87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.906074] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129806, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03217} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.906074] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.906074] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.906074] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk to [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.906624] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07d6bec8-c255-4037-9b36-fd699f812f8d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.912862] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 835.912862] env[61991]: value = "task-1129807" [ 835.912862] env[61991]: _type = "Task" [ 835.912862] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.921331] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.008114] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.008114] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.010454] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.581s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.011914] env[61991]: INFO nova.compute.claims [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.381258] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52099bf4-6497-4b8b-9ef6-c0a4b55da63a, 'name': SearchDatastore_Task, 'duration_secs': 0.009662} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.381555] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.381856] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.382165] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39e4c6b1-734b-4235-b371-0bff1e075cff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.387414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.387414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.387414] env[61991]: DEBUG nova.network.neutron [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.391286] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 836.391286] env[61991]: value = "task-1129808" [ 836.391286] env[61991]: _type = "Task" [ 836.391286] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.404790] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.424491] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129807, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.519497] env[61991]: DEBUG nova.compute.utils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.520694] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.520871] env[61991]: DEBUG nova.network.neutron [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 836.559365] env[61991]: DEBUG nova.policy [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '160b84053b8a4d7c86b013e751db6b66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a2b37cc1ade437097583e823d457ca6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.843369] env[61991]: DEBUG nova.network.neutron [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Successfully created port: c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.903787] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.927261] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129807, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.946444] env[61991]: DEBUG nova.network.neutron [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 837.029310] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.241913] env[61991]: DEBUG nova.network.neutron [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updating instance_info_cache with network_info: [{"id": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "address": "fa:16:3e:81:55:e0", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e2f32e-53", "ovs_interfaceid": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.366904] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef0193d-98e7-47e2-ab21-10bdfb0e15f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.377327] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78846d7-848d-410e-a388-933ccdb5e257 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.416577] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bee612-d95a-4461-a1fe-bedd392ee029 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.427672] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129807, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.432579] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.433973] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee50354d-7319-4e58-9f93-6f5891d269a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.450128] env[61991]: DEBUG nova.compute.provider_tree [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.747130] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.747572] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Instance network_info: |[{"id": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "address": "fa:16:3e:81:55:e0", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e2f32e-53", "ovs_interfaceid": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.748101] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:55:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0df968ae-c1ef-4009-a0f4-6f2e799c2fda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87e2f32e-53df-49a9-bae3-f9529f2d3c6c', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.755858] env[61991]: DEBUG oslo.service.loopingcall [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.756170] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.756427] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2905e4ef-6826-4868-abe0-e5704b625e86 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.776908] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.776908] env[61991]: value = "task-1129809" [ 837.776908] env[61991]: _type = "Task" [ 837.776908] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.787752] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129809, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.825962] env[61991]: DEBUG nova.compute.manager [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Received event network-changed-87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.826194] env[61991]: DEBUG nova.compute.manager [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Refreshing instance network info cache due to event network-changed-87e2f32e-53df-49a9-bae3-f9529f2d3c6c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.826428] env[61991]: DEBUG oslo_concurrency.lockutils [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] Acquiring lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.826588] env[61991]: DEBUG oslo_concurrency.lockutils [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] Acquired lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.826753] env[61991]: DEBUG nova.network.neutron [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Refreshing network info cache for port 87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.927774] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129808, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.934983] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129807, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.953225] env[61991]: DEBUG nova.scheduler.client.report [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.037051] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.063305] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.063638] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.063844] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.064084] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.064302] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.064537] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.064780] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.065022] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.065261] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.065476] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.065718] env[61991]: DEBUG nova.virt.hardware [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.066820] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58112370-14c5-453c-9f23-7d3cde48d96a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.078031] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b07627f-2769-4ef0-bf8e-f168993f2fa3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.288381] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129809, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.427449] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129808, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.434266] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129807, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.427803} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.434266] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0810354d-ca0b-4776-8487-42797b713043/0810354d-ca0b-4776-8487-42797b713043.vmdk to [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.434266] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdf168f-75eb-4cb6-9237-fd9225eb7c9a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.464179] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.465052] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.465580] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.468304] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1e04eac-007b-4f47-8191-bba3269d96bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.486685] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.517s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.487733] env[61991]: DEBUG nova.objects.instance [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lazy-loading 'resources' on Instance uuid 1715711a-72c5-4e86-88e0-cecf922cb42f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.496803] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 838.496803] env[61991]: value = "task-1129810" [ 838.496803] env[61991]: _type = "Task" [ 838.496803] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.508234] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129810, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.775210] env[61991]: DEBUG nova.compute.manager [req-dd924a68-1d9f-49c7-a2ae-a44cd5c1b728 req-5c57a8c4-8dda-4a59-8e67-a68b70546d6e service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-vif-plugged-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.775450] env[61991]: DEBUG oslo_concurrency.lockutils [req-dd924a68-1d9f-49c7-a2ae-a44cd5c1b728 req-5c57a8c4-8dda-4a59-8e67-a68b70546d6e service nova] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.775655] env[61991]: DEBUG oslo_concurrency.lockutils [req-dd924a68-1d9f-49c7-a2ae-a44cd5c1b728 req-5c57a8c4-8dda-4a59-8e67-a68b70546d6e service nova] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.775873] env[61991]: DEBUG oslo_concurrency.lockutils [req-dd924a68-1d9f-49c7-a2ae-a44cd5c1b728 req-5c57a8c4-8dda-4a59-8e67-a68b70546d6e service nova] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.775998] env[61991]: DEBUG nova.compute.manager [req-dd924a68-1d9f-49c7-a2ae-a44cd5c1b728 req-5c57a8c4-8dda-4a59-8e67-a68b70546d6e service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] No waiting events found dispatching network-vif-plugged-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 838.776252] env[61991]: WARNING nova.compute.manager [req-dd924a68-1d9f-49c7-a2ae-a44cd5c1b728 req-5c57a8c4-8dda-4a59-8e67-a68b70546d6e service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received unexpected event network-vif-plugged-c7263e46-a164-4813-be3e-faae7fc31ff2 for instance with vm_state building and task_state spawning. [ 838.788279] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129809, 'name': CreateVM_Task, 'duration_secs': 0.870574} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.788440] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.789717] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.789717] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.789717] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.789832] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48a884db-8fbb-4d67-a0a0-3ed5de6dd2d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.793923] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 838.793923] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5206e28b-2b8d-6cb5-ebac-bcbdffde9a5e" [ 838.793923] env[61991]: _type = "Task" [ 838.793923] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.806561] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5206e28b-2b8d-6cb5-ebac-bcbdffde9a5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.831243] env[61991]: DEBUG nova.network.neutron [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Successfully updated port: c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.880903] env[61991]: DEBUG nova.network.neutron [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updated VIF entry in instance network info cache for port 87e2f32e-53df-49a9-bae3-f9529f2d3c6c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.881308] env[61991]: DEBUG nova.network.neutron [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updating instance_info_cache with network_info: [{"id": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "address": "fa:16:3e:81:55:e0", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e2f32e-53", "ovs_interfaceid": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.923563] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129808, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.037728} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.923803] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.924027] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.925116] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ccfb93da-85b8-4326-b420-a333a5f2e41f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.931368] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 838.931368] env[61991]: value = "task-1129811" [ 838.931368] env[61991]: _type = "Task" [ 838.931368] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.939138] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.986273] env[61991]: DEBUG nova.compute.utils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.987828] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.987998] env[61991]: DEBUG nova.network.neutron [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 839.009181] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129810, 'name': ReconfigVM_Task, 'duration_secs': 0.271797} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.009489] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Reconfigured VM instance instance-00000031 to attach disk [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e/f78ef63b-453e-45d3-959b-4b0c1922b53e.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.010111] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6e9f857-ead4-4cd6-bea9-c5690379d5a6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.015788] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 839.015788] env[61991]: value = "task-1129812" [ 839.015788] env[61991]: _type = "Task" [ 839.015788] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.023842] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129812, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.031225] env[61991]: DEBUG nova.policy [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a385cf261940a19eb17046958920ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3015571d285418aae0fca50b563394e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.260445] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c717b02-02ad-493a-9d7e-29d68a1ea425 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.268578] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8590241f-57ba-4df8-9e21-2ce9b1a1c97d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.299475] env[61991]: DEBUG nova.network.neutron [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Successfully created port: 84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.304253] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483fbf17-b148-4851-8e05-07dd5649cd50 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.312082] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5206e28b-2b8d-6cb5-ebac-bcbdffde9a5e, 'name': SearchDatastore_Task, 'duration_secs': 0.008967} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.314052] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.314295] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.314525] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.314673] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.314850] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.315170] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c4e86ca-d43a-4c85-98c6-fd5bc22c0be4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.317659] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ff99bd-994b-4a8e-afcc-e9283f51155c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.331251] env[61991]: DEBUG nova.compute.provider_tree [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.333333] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.333507] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.334396] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ee21b2f-9492-430e-ac35-22fcfad79e75 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.336810] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.336939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.337093] env[61991]: DEBUG nova.network.neutron [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.340821] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 839.340821] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520d3359-2a01-b7dd-8da8-9c671e9484f9" [ 839.340821] env[61991]: _type = "Task" [ 839.340821] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.352924] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520d3359-2a01-b7dd-8da8-9c671e9484f9, 'name': SearchDatastore_Task, 'duration_secs': 0.008677} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.353676] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d08e2d1-11cd-438f-ac33-7e4a57afcde3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.358753] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 839.358753] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5266ffbe-72ec-60f2-ba5f-fec913b130ef" [ 839.358753] env[61991]: _type = "Task" [ 839.358753] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.367818] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5266ffbe-72ec-60f2-ba5f-fec913b130ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.383460] env[61991]: DEBUG oslo_concurrency.lockutils [req-5088336b-e7b9-4f5f-b417-2b7e78c0c4dc req-bfd16331-8bd4-49fe-bad5-5b7315ea7e94 service nova] Releasing lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.441100] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062218} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.441378] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.442201] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7c3138-e881-467f-a8aa-82531ead0749 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.460903] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.461913] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91c09f30-5c60-49c7-84a3-b4ec626098e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.479774] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 839.479774] env[61991]: value = "task-1129813" [ 839.479774] env[61991]: _type = "Task" [ 839.479774] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.490018] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129813, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.490661] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.526171] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129812, 'name': Rename_Task, 'duration_secs': 0.135876} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.526433] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.526628] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-301c943e-4f33-42cb-a781-88fb118c0277 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.533349] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 839.533349] env[61991]: value = "task-1129814" [ 839.533349] env[61991]: _type = "Task" [ 839.533349] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.542435] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129814, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.836014] env[61991]: DEBUG nova.scheduler.client.report [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.868538] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5266ffbe-72ec-60f2-ba5f-fec913b130ef, 'name': SearchDatastore_Task, 'duration_secs': 0.008462} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.868772] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.869049] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/c1d4d2eb-810c-42f9-a244-e5aed6e8d517.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.869309] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5dd66d9a-cda0-4cbb-9077-90248e4415cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.875509] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 839.875509] env[61991]: value = "task-1129815" [ 839.875509] env[61991]: _type = "Task" [ 839.875509] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.882926] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129815, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.886835] env[61991]: DEBUG nova.compute.manager [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-changed-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.887110] env[61991]: DEBUG nova.compute.manager [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Refreshing instance network info cache due to event network-changed-c7263e46-a164-4813-be3e-faae7fc31ff2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.887250] env[61991]: DEBUG oslo_concurrency.lockutils [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.888069] env[61991]: DEBUG nova.network.neutron [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 839.990466] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129813, 'name': ReconfigVM_Task, 'duration_secs': 0.290463} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.990868] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.991475] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eeaa12ea-6271-4edd-9ddd-bd4675d8fe9d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.999166] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 839.999166] env[61991]: value = "task-1129816" [ 839.999166] env[61991]: _type = "Task" [ 839.999166] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.007428] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129816, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.045946] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129814, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.051837] env[61991]: DEBUG nova.network.neutron [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.342133] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.346093] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.767s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.346093] env[61991]: DEBUG nova.objects.instance [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 840.367476] env[61991]: INFO nova.scheduler.client.report [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted allocations for instance 1715711a-72c5-4e86-88e0-cecf922cb42f [ 840.386804] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129815, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480198} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.387405] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/c1d4d2eb-810c-42f9-a244-e5aed6e8d517.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.387405] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.387616] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe795813-2e7b-4a05-a090-5e46a31c1fe3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.394142] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 840.394142] env[61991]: value = "task-1129817" [ 840.394142] env[61991]: _type = "Task" [ 840.394142] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.402201] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129817, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.500761] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.511458] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129816, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.524453] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.524692] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.524905] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.525226] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.525392] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.525544] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.525751] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.525912] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.526112] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.526286] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.526461] env[61991]: DEBUG nova.virt.hardware [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.527237] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29ab79b-a385-4b87-8e3c-c0c2696c8342 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.534833] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06970246-5dfd-4705-b91d-a10348eb8f2d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.552631] env[61991]: DEBUG oslo_vmware.api [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129814, 'name': PowerOnVM_Task, 'duration_secs': 0.670654} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.552708] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.555008] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.555315] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance network_info: |[{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.555591] env[61991]: DEBUG oslo_concurrency.lockutils [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.555746] env[61991]: DEBUG nova.network.neutron [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Refreshing network info cache for port c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 840.556875] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:d4:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7263e46-a164-4813-be3e-faae7fc31ff2', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.564983] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating folder: Project (9a2b37cc1ade437097583e823d457ca6). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.565926] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad8f8278-7b64-4523-8e41-3ccd46189e96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.578237] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Created folder: Project (9a2b37cc1ade437097583e823d457ca6) in parent group-v246753. [ 840.578237] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating folder: Instances. Parent ref: group-v246889. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.578237] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83a2be84-3f10-4249-aa32-9b3e0d66cf38 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.586697] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Created folder: Instances in parent group-v246889. [ 840.587316] env[61991]: DEBUG oslo.service.loopingcall [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.587316] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.587316] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54c616b1-905d-4aca-9cfd-a999d528a556 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.605641] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.605641] env[61991]: value = "task-1129820" [ 840.605641] env[61991]: _type = "Task" [ 840.605641] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.612654] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129820, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.658410] env[61991]: DEBUG nova.compute.manager [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.659328] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97fe561-2796-41d8-80d8-1952f75ed0c9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.877112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e783689e-2e12-4666-9527-680f66f7bf9f tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "1715711a-72c5-4e86-88e0-cecf922cb42f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.315s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.905128] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129817, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063415} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.905419] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.906225] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2024930d-9545-4202-b370-00549a02f10f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.930473] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/c1d4d2eb-810c-42f9-a244-e5aed6e8d517.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.931289] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7889eeb0-ce34-47d7-8d09-0aaf2faf05d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.951028] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 840.951028] env[61991]: value = "task-1129821" [ 840.951028] env[61991]: _type = "Task" [ 840.951028] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.958852] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129821, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.966727] env[61991]: DEBUG nova.network.neutron [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Successfully updated port: 84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.013691] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129816, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.115080] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129820, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.176104] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f5e6cb72-185e-4487-a877-83f819facb50 tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 33.387s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.260561] env[61991]: DEBUG nova.network.neutron [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updated VIF entry in instance network info cache for port c7263e46-a164-4813-be3e-faae7fc31ff2. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 841.260948] env[61991]: DEBUG nova.network.neutron [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.357985] env[61991]: DEBUG oslo_concurrency.lockutils [None req-16cf0b96-7fc8-4c6f-9ae7-b2dcf1c6dde1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.359221] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.474s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.359737] env[61991]: INFO nova.compute.claims [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.462601] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129821, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.469710] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "refresh_cache-7a468ab7-6ca1-43e2-a9b2-89e7640d8148" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.469850] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "refresh_cache-7a468ab7-6ca1-43e2-a9b2-89e7640d8148" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.470010] env[61991]: DEBUG nova.network.neutron [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.514290] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129816, 'name': Rename_Task, 'duration_secs': 1.142183} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.514693] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.514988] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16cedbf1-57a3-4573-8662-7c1d09eb4cd7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.521121] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 841.521121] env[61991]: value = "task-1129822" [ 841.521121] env[61991]: _type = "Task" [ 841.521121] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.528696] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129822, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.617891] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129820, 'name': CreateVM_Task, 'duration_secs': 0.917745} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.618091] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 841.618738] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.618912] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.619454] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.619614] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34f844d4-3d75-4062-b390-9cb46790237b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.624101] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 841.624101] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527bb481-74bd-b9bc-64b8-ec66cce8fbfd" [ 841.624101] env[61991]: _type = "Task" [ 841.624101] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.631613] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527bb481-74bd-b9bc-64b8-ec66cce8fbfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.763392] env[61991]: DEBUG oslo_concurrency.lockutils [req-acbea0c6-69d1-4468-8e94-1706d65c8678 req-ae2722ac-5d02-4f50-ab3a-9a524aa9cd60 service nova] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.924624] env[61991]: DEBUG nova.compute.manager [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Received event network-vif-plugged-84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.924946] env[61991]: DEBUG oslo_concurrency.lockutils [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] Acquiring lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.925289] env[61991]: DEBUG oslo_concurrency.lockutils [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.925489] env[61991]: DEBUG oslo_concurrency.lockutils [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.925671] env[61991]: DEBUG nova.compute.manager [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] No waiting events found dispatching network-vif-plugged-84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.925840] env[61991]: WARNING nova.compute.manager [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Received unexpected event network-vif-plugged-84517dfd-59cc-45de-bde0-638279ae93c8 for instance with vm_state building and task_state spawning. [ 841.926051] env[61991]: DEBUG nova.compute.manager [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Received event network-changed-84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.926290] env[61991]: DEBUG nova.compute.manager [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Refreshing instance network info cache due to event network-changed-84517dfd-59cc-45de-bde0-638279ae93c8. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 841.926484] env[61991]: DEBUG oslo_concurrency.lockutils [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] Acquiring lock "refresh_cache-7a468ab7-6ca1-43e2-a9b2-89e7640d8148" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.961780] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129821, 'name': ReconfigVM_Task, 'duration_secs': 0.628851} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.962079] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Reconfigured VM instance instance-00000049 to attach disk [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/c1d4d2eb-810c-42f9-a244-e5aed6e8d517.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.962716] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99a040cf-d406-401f-83c4-5f9d1ef10865 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.969913] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 841.969913] env[61991]: value = "task-1129823" [ 841.969913] env[61991]: _type = "Task" [ 841.969913] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.980793] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129823, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.018442] env[61991]: DEBUG nova.network.neutron [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.030256] env[61991]: DEBUG oslo_vmware.api [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129822, 'name': PowerOnVM_Task, 'duration_secs': 0.461268} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.030532] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.030744] env[61991]: INFO nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Took 8.87 seconds to spawn the instance on the hypervisor. [ 842.030924] env[61991]: DEBUG nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.031736] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeda3dc5-5631-448f-a4f7-d99a1796b061 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.134527] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527bb481-74bd-b9bc-64b8-ec66cce8fbfd, 'name': SearchDatastore_Task, 'duration_secs': 0.008286} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.134801] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.135106] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.135362] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.135515] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.135695] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.135976] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa9f9731-8167-43c9-bae5-53532bb3261b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.144899] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.144899] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 842.147036] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c5dc7da-3695-42a0-963c-8af1b7b9c10d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.151622] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 842.151622] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52323b76-fa51-a6ae-1cd6-bd12a5f7cf2a" [ 842.151622] env[61991]: _type = "Task" [ 842.151622] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.158960] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52323b76-fa51-a6ae-1cd6-bd12a5f7cf2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.205349] env[61991]: DEBUG nova.network.neutron [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Updating instance_info_cache with network_info: [{"id": "84517dfd-59cc-45de-bde0-638279ae93c8", "address": "fa:16:3e:8c:00:65", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84517dfd-59", "ovs_interfaceid": "84517dfd-59cc-45de-bde0-638279ae93c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.482010] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129823, 'name': Rename_Task, 'duration_secs': 0.131976} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.482325] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.482569] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45354063-85d7-41ba-a0ca-7a04fd199e7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.488944] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 842.488944] env[61991]: value = "task-1129824" [ 842.488944] env[61991]: _type = "Task" [ 842.488944] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.496078] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129824, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.550215] env[61991]: INFO nova.compute.manager [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Took 28.45 seconds to build instance. [ 842.617184] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cec0a5a-fce9-4072-b01a-503ede3e0d92 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.626390] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e9bc59-ed98-4f2e-899d-8943a06ef7ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.658783] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40548a7-c7be-40c7-9df4-67cb3b915181 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.666680] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52323b76-fa51-a6ae-1cd6-bd12a5f7cf2a, 'name': SearchDatastore_Task, 'duration_secs': 0.007862} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.669110] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaca84ee-ba92-4c29-8609-1c1a3972a0fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.672644] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be47be0f-0663-4721-8928-5889da38a523 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.679327] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 842.679327] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521897cf-71e5-ca8a-af79-7862eb1168f9" [ 842.679327] env[61991]: _type = "Task" [ 842.679327] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.687436] env[61991]: DEBUG nova.compute.provider_tree [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.696226] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521897cf-71e5-ca8a-af79-7862eb1168f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.708143] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "refresh_cache-7a468ab7-6ca1-43e2-a9b2-89e7640d8148" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.708455] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Instance network_info: |[{"id": "84517dfd-59cc-45de-bde0-638279ae93c8", "address": "fa:16:3e:8c:00:65", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84517dfd-59", "ovs_interfaceid": "84517dfd-59cc-45de-bde0-638279ae93c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.708724] env[61991]: DEBUG oslo_concurrency.lockutils [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] Acquired lock "refresh_cache-7a468ab7-6ca1-43e2-a9b2-89e7640d8148" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.708902] env[61991]: DEBUG nova.network.neutron [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Refreshing network info cache for port 84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 842.710027] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:00:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3952eb02-1162-48ed-8227-9c138960d583', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84517dfd-59cc-45de-bde0-638279ae93c8', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.717475] env[61991]: DEBUG oslo.service.loopingcall [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.720476] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.720912] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ec1b31a-6b80-4d41-be58-a57b04a31c7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.740081] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.740081] env[61991]: value = "task-1129825" [ 842.740081] env[61991]: _type = "Task" [ 842.740081] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.747983] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129825, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.000635] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129824, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.011810] env[61991]: DEBUG nova.network.neutron [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Updated VIF entry in instance network info cache for port 84517dfd-59cc-45de-bde0-638279ae93c8. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.012249] env[61991]: DEBUG nova.network.neutron [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Updating instance_info_cache with network_info: [{"id": "84517dfd-59cc-45de-bde0-638279ae93c8", "address": "fa:16:3e:8c:00:65", "network": {"id": "b2a899ea-d3ea-4c28-8394-1c1f6493e5d1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-186062868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3015571d285418aae0fca50b563394e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3952eb02-1162-48ed-8227-9c138960d583", "external-id": "nsx-vlan-transportzone-250", "segmentation_id": 250, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84517dfd-59", "ovs_interfaceid": "84517dfd-59cc-45de-bde0-638279ae93c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.052110] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1594ad78-0ab4-40db-a13a-5384ca93e066 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "96627657-d3d9-4a64-b83e-87323bd10c03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.967s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.199022] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521897cf-71e5-ca8a-af79-7862eb1168f9, 'name': SearchDatastore_Task, 'duration_secs': 0.022207} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.199365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.199698] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 843.200014] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2572a0d5-75bb-4640-aa14-23e860f80053 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.206696] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 843.206696] env[61991]: value = "task-1129826" [ 843.206696] env[61991]: _type = "Task" [ 843.206696] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.215155] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129826, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.217242] env[61991]: ERROR nova.scheduler.client.report [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [req-cdac3173-e478-451f-9937-bf162958d6c9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cdac3173-e478-451f-9937-bf162958d6c9"}]} [ 843.240301] env[61991]: DEBUG nova.scheduler.client.report [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 843.251423] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129825, 'name': CreateVM_Task, 'duration_secs': 0.366917} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.251626] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.252394] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.252563] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.252894] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.253520] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cee6df8f-5702-4f7e-abc8-f1d051ce438f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.255980] env[61991]: DEBUG nova.scheduler.client.report [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 843.256246] env[61991]: DEBUG nova.compute.provider_tree [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.261405] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 843.261405] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ca2069-4263-223a-f460-f32d4f242083" [ 843.261405] env[61991]: _type = "Task" [ 843.261405] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.270382] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ca2069-4263-223a-f460-f32d4f242083, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.271319] env[61991]: DEBUG nova.scheduler.client.report [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 843.293820] env[61991]: DEBUG nova.scheduler.client.report [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 843.499233] env[61991]: DEBUG oslo_vmware.api [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129824, 'name': PowerOnVM_Task, 'duration_secs': 0.802073} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.501609] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.501822] env[61991]: INFO nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Took 7.95 seconds to spawn the instance on the hypervisor. [ 843.502009] env[61991]: DEBUG nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.502948] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef6c987-62b7-4b1f-b247-01183a067c23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.514912] env[61991]: DEBUG oslo_concurrency.lockutils [req-8a75ec03-2370-4464-9284-a8514ebfa25a req-3bbb30b4-ca54-4a86-8c38-0f389eb96383 service nova] Releasing lock "refresh_cache-7a468ab7-6ca1-43e2-a9b2-89e7640d8148" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.534015] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7378c482-4acb-4a37-987b-f7169251d92f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.541877] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408c8070-63fa-4231-981d-c6dd24dede77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.574192] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.577557] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96577437-5db1-4b7d-abf0-49f59ab65e06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.585810] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df49193a-c907-4caf-b022-1a1bce5c2a8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.600478] env[61991]: DEBUG nova.compute.provider_tree [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.721776] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129826, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.723508] env[61991]: INFO nova.compute.manager [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Rebuilding instance [ 843.768826] env[61991]: DEBUG nova.compute.manager [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.769712] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af2ac04-a045-48e0-9b0b-3d3b1cd78585 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.776197] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ca2069-4263-223a-f460-f32d4f242083, 'name': SearchDatastore_Task, 'duration_secs': 0.010324} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.776927] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.777191] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.777453] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.777596] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.777775] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.778059] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a395523-7f53-4573-bb5c-f32cda1fafd3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.792837] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.793086] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.793950] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1bb4e80-a12d-4e7d-b326-618871820c91 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.800946] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 843.800946] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b4294c-e022-5dbc-189d-0f67db00d353" [ 843.800946] env[61991]: _type = "Task" [ 843.800946] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.810901] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b4294c-e022-5dbc-189d-0f67db00d353, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.881603] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 843.881949] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.020896] env[61991]: INFO nova.compute.manager [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Took 29.85 seconds to build instance. [ 844.097823] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.139411] env[61991]: DEBUG nova.scheduler.client.report [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 844.139844] env[61991]: DEBUG nova.compute.provider_tree [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 99 to 100 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 844.140165] env[61991]: DEBUG nova.compute.provider_tree [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.218728] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129826, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.831615} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.218963] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 844.219229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.219490] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0342baa5-5cf4-423a-bfe9-4a0d49955aa8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.226523] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 844.226523] env[61991]: value = "task-1129827" [ 844.226523] env[61991]: _type = "Task" [ 844.226523] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.234438] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129827, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.286183] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.286551] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-173c60b4-c9f1-4bd8-9eaf-486d2453b8fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.294213] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 844.294213] env[61991]: value = "task-1129828" [ 844.294213] env[61991]: _type = "Task" [ 844.294213] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.302561] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.310530] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b4294c-e022-5dbc-189d-0f67db00d353, 'name': SearchDatastore_Task, 'duration_secs': 0.022226} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.311298] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ff5de7d-816f-4abf-bb70-a716d2ea2b6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.316454] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 844.316454] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280ff3f-b6c6-0f47-35cb-8062fa2a1be9" [ 844.316454] env[61991]: _type = "Task" [ 844.316454] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.326857] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280ff3f-b6c6-0f47-35cb-8062fa2a1be9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.388642] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 844.388881] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 844.523328] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9d87e821-e92a-4a08-bf05-31665d2ef6aa tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.366s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.646579] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.288s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.647136] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.649973] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.668s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.650254] env[61991]: DEBUG nova.objects.instance [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'resources' on Instance uuid 49324fef-ad48-451b-a5ce-d9a1231137db {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.736579] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129827, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.212046} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.736868] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.737700] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c5f4e8-2128-4b33-8d79-532bc1940bac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.758951] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.759150] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aeb21709-baf7-4a2e-a2cd-b61eb3962fbf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.777636] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 844.777636] env[61991]: value = "task-1129829" [ 844.777636] env[61991]: _type = "Task" [ 844.777636] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.785143] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129829, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.802913] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129828, 'name': PowerOffVM_Task, 'duration_secs': 0.227709} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.803453] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 844.803694] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.804452] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2407f67-8935-4b4d-b7f6-1f16b9c0cd13 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.810551] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 844.810765] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38b6c895-c5bf-4e79-a3dc-6cb6b3d5e3d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.825362] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280ff3f-b6c6-0f47-35cb-8062fa2a1be9, 'name': SearchDatastore_Task, 'duration_secs': 0.036962} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.825615] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.826090] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 7a468ab7-6ca1-43e2-a9b2-89e7640d8148/7a468ab7-6ca1-43e2-a9b2-89e7640d8148.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.826336] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-535f0fb2-9c62-4ce8-aec4-b76e6a713f7b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.832451] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 844.832451] env[61991]: value = "task-1129831" [ 844.832451] env[61991]: _type = "Task" [ 844.832451] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.835988] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 844.836224] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 844.836413] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Deleting the datastore file [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.836913] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32067ae5-ec7f-4afb-8a79-719ca505c9ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.841141] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.845600] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 844.845600] env[61991]: value = "task-1129832" [ 844.845600] env[61991]: _type = "Task" [ 844.845600] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.852513] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129832, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.091173] env[61991]: INFO nova.compute.manager [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Rescuing [ 845.091483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.091643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.091814] env[61991]: DEBUG nova.network.neutron [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.152822] env[61991]: DEBUG nova.compute.utils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.154399] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.154573] env[61991]: DEBUG nova.network.neutron [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 845.243566] env[61991]: DEBUG nova.policy [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6422dfa4c424fbaab78c3e72a0e6669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dddb51029854105bc6c9c8724181d39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.289860] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129829, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.346469] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.357081] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129832, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.488791} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.357387] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.357591] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 845.357778] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.460541] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06135108-9793-43a3-88f0-2d13e3b02a1b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.468897] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66066046-07cb-482d-80be-eb805326f633 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.499799] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d879714d-f349-4602-a875-1d50aa8b9e3d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.507494] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddb0ccc-e506-404e-a755-4d8d4fefc1a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.522596] env[61991]: DEBUG nova.compute.provider_tree [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 845.598852] env[61991]: DEBUG nova.network.neutron [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Successfully created port: 58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.659748] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.788247] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129829, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.846544] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.906826} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.846810] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 7a468ab7-6ca1-43e2-a9b2-89e7640d8148/7a468ab7-6ca1-43e2-a9b2-89e7640d8148.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.847042] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.847316] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15fad3b9-77f2-4902-8216-76b1eb083e18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.853721] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 845.853721] env[61991]: value = "task-1129833" [ 845.853721] env[61991]: _type = "Task" [ 845.853721] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.864053] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129833, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.993910] env[61991]: DEBUG nova.network.neutron [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updating instance_info_cache with network_info: [{"id": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "address": "fa:16:3e:81:55:e0", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e2f32e-53", "ovs_interfaceid": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.057869] env[61991]: DEBUG nova.scheduler.client.report [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 100 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 846.058188] env[61991]: DEBUG nova.compute.provider_tree [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 100 to 101 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 846.058379] env[61991]: DEBUG nova.compute.provider_tree [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 846.288986] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129829, 'name': ReconfigVM_Task, 'duration_secs': 1.337655} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.289292] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.289980] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-174137e5-c423-4c7e-9fff-a122a3f76a07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.296123] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 846.296123] env[61991]: value = "task-1129834" [ 846.296123] env[61991]: _type = "Task" [ 846.296123] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.303652] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129834, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.363730] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129833, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0623} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.363730] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.364496] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c966744-74a8-429c-ae78-335c8a27b8b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.389706] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 7a468ab7-6ca1-43e2-a9b2-89e7640d8148/7a468ab7-6ca1-43e2-a9b2-89e7640d8148.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.391962] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a506bf7c-8eb9-4624-b749-484096e6b00f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.413641] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.413794] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.414469] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.414469] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.414469] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.414469] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.414788] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.414853] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.414988] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.415633] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.415633] env[61991]: DEBUG nova.virt.hardware [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.416293] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c24654-47e0-45cf-ac81-5a3615352d1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.420343] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 846.420343] env[61991]: value = "task-1129835" [ 846.420343] env[61991]: _type = "Task" [ 846.420343] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.427071] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51665be4-077f-4e9c-a8bf-13230e836dd6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.433965] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129835, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.444651] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.450227] env[61991]: DEBUG oslo.service.loopingcall [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.451275] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.451361] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.451423] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 846.454030] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.454030] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d593f168-246a-4e01-9305-a0f79b6538f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.473710] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.473710] env[61991]: value = "task-1129836" [ 846.473710] env[61991]: _type = "Task" [ 846.473710] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.482510] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129836, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.497200] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.563728] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.566244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.871s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.567843] env[61991]: INFO nova.compute.claims [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.587872] env[61991]: INFO nova.scheduler.client.report [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted allocations for instance 49324fef-ad48-451b-a5ce-d9a1231137db [ 846.669946] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.692895] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.693214] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.693427] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.693598] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.693931] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.694195] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.694389] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.694514] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.694809] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.695057] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.695283] env[61991]: DEBUG nova.virt.hardware [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.696172] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f1ffdb-e1c9-4559-9b08-e070f148cbdf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.703723] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb155701-8c50-4c1e-9154-443d0717ef1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.807840] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129834, 'name': Rename_Task, 'duration_secs': 0.141869} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.808142] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.808391] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b974d5b-e490-4b2e-997a-ef787034aeed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.815790] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 846.815790] env[61991]: value = "task-1129837" [ 846.815790] env[61991]: _type = "Task" [ 846.815790] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.823459] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129837, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.930838] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129835, 'name': ReconfigVM_Task, 'duration_secs': 0.29871} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.931198] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 7a468ab7-6ca1-43e2-a9b2-89e7640d8148/7a468ab7-6ca1-43e2-a9b2-89e7640d8148.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.931861] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-265b0f9c-871a-4996-9d56-5523c8606cba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.938342] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 846.938342] env[61991]: value = "task-1129838" [ 846.938342] env[61991]: _type = "Task" [ 846.938342] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.946673] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129838, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.983020] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129836, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.026848] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.027253] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96392049-c6af-48ac-922a-08fd34261455 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.035183] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 847.035183] env[61991]: value = "task-1129839" [ 847.035183] env[61991]: _type = "Task" [ 847.035183] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.045749] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.100310] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a6bad93d-529e-4817-9329-c3527c6cdc51 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "49324fef-ad48-451b-a5ce-d9a1231137db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.498s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.281354] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.281581] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.281765] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.281952] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.282148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.284374] env[61991]: INFO nova.compute.manager [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Terminating instance [ 847.286319] env[61991]: DEBUG nova.compute.manager [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.286477] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 847.287228] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad90aa5-cbac-4ae3-a820-c3cf57bf2ad8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.294755] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.295090] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b147190-9f4b-48c1-acab-e8e0bfc338d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.300721] env[61991]: DEBUG oslo_vmware.api [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 847.300721] env[61991]: value = "task-1129840" [ 847.300721] env[61991]: _type = "Task" [ 847.300721] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.307881] env[61991]: DEBUG oslo_vmware.api [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129840, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.321606] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating instance_info_cache with network_info: [{"id": "87a47717-3b2c-4160-97b7-ef566ac540c0", "address": "fa:16:3e:e6:5f:4d", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.180", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87a47717-3b", "ovs_interfaceid": "87a47717-3b2c-4160-97b7-ef566ac540c0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.326152] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129837, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.337886] env[61991]: DEBUG nova.compute.manager [req-55a85eda-77d1-4a5e-a99a-6b00eb8d7090 req-c6292ccb-d846-45b7-bf62-c8900a65d8a0 service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Received event network-vif-plugged-58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.338141] env[61991]: DEBUG oslo_concurrency.lockutils [req-55a85eda-77d1-4a5e-a99a-6b00eb8d7090 req-c6292ccb-d846-45b7-bf62-c8900a65d8a0 service nova] Acquiring lock "3183190e-3202-40be-894c-e47b186708b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.338358] env[61991]: DEBUG oslo_concurrency.lockutils [req-55a85eda-77d1-4a5e-a99a-6b00eb8d7090 req-c6292ccb-d846-45b7-bf62-c8900a65d8a0 service nova] Lock "3183190e-3202-40be-894c-e47b186708b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.338529] env[61991]: DEBUG oslo_concurrency.lockutils [req-55a85eda-77d1-4a5e-a99a-6b00eb8d7090 req-c6292ccb-d846-45b7-bf62-c8900a65d8a0 service nova] Lock "3183190e-3202-40be-894c-e47b186708b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.338703] env[61991]: DEBUG nova.compute.manager [req-55a85eda-77d1-4a5e-a99a-6b00eb8d7090 req-c6292ccb-d846-45b7-bf62-c8900a65d8a0 service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] No waiting events found dispatching network-vif-plugged-58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.338871] env[61991]: WARNING nova.compute.manager [req-55a85eda-77d1-4a5e-a99a-6b00eb8d7090 req-c6292ccb-d846-45b7-bf62-c8900a65d8a0 service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Received unexpected event network-vif-plugged-58dc3a16-cc46-440b-b310-026854eda3a2 for instance with vm_state building and task_state spawning. [ 847.431362] env[61991]: DEBUG nova.network.neutron [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Successfully updated port: 58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.448393] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129838, 'name': Rename_Task, 'duration_secs': 0.172381} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.448687] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.448933] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05a03d71-0c96-4f77-a7e8-78fa2f27bb5d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.456169] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 847.456169] env[61991]: value = "task-1129841" [ 847.456169] env[61991]: _type = "Task" [ 847.456169] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.463910] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129841, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.483611] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129836, 'name': CreateVM_Task, 'duration_secs': 0.517417} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.483795] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.484258] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.484434] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.484776] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.485045] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0497f921-eccf-46ef-9ce6-7e3c4ce58d93 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.490190] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 847.490190] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526399a2-461c-3740-a00d-54cc781a1692" [ 847.490190] env[61991]: _type = "Task" [ 847.490190] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.500398] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526399a2-461c-3740-a00d-54cc781a1692, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.543975] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129839, 'name': PowerOffVM_Task, 'duration_secs': 0.19103} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.544276] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.545090] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277b91a5-a3fc-41c6-a1de-c1c67276bda4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.563659] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe526fa3-a928-4dcf-910a-67ea84d8ea58 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.607217] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.607584] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41394ba1-3e2b-4f2f-b20b-6a181122a713 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.617557] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 847.617557] env[61991]: value = "task-1129842" [ 847.617557] env[61991]: _type = "Task" [ 847.617557] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.630904] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 847.631189] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.631450] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.631674] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.631912] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.632524] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f79d294-4bec-4c46-9c28-daf4ad77a136 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.641548] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.641809] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.648546] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2471406-9cad-4518-85d6-c5ebb1629dbd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.658446] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 847.658446] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5246de71-1f78-783f-8129-65018f51f896" [ 847.658446] env[61991]: _type = "Task" [ 847.658446] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.666489] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5246de71-1f78-783f-8129-65018f51f896, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.810983] env[61991]: DEBUG oslo_vmware.api [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129840, 'name': PowerOffVM_Task, 'duration_secs': 0.370093} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.813473] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.813655] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.814099] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55f1179f-192d-4948-bda0-51c362ee8026 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.825641] env[61991]: DEBUG oslo_vmware.api [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129837, 'name': PowerOnVM_Task, 'duration_secs': 0.868854} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.825899] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.826161] env[61991]: INFO nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Took 9.79 seconds to spawn the instance on the hypervisor. [ 847.826561] env[61991]: DEBUG nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.826912] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-9c17f844-1f0e-4f01-aadc-0f1f75a59d06" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.827174] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 847.827908] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375aaa44-d774-43ec-90e4-46caa205f904 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.832691] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.833158] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.833781] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.834061] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.834324] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.834585] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.834724] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 847.834908] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 847.882965] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.883262] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.883455] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleting the datastore file [datastore2] 62b3e5ae-b90c-47f8-95c4-14587dbf647d {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.883719] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e5072dd-06f0-4bcf-804a-6941a3e8b932 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.890424] env[61991]: DEBUG oslo_vmware.api [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 847.890424] env[61991]: value = "task-1129844" [ 847.890424] env[61991]: _type = "Task" [ 847.890424] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.896522] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6144bcd4-7e9b-476c-b423-68190842bc9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.907023] env[61991]: DEBUG oslo_vmware.api [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.907336] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebafa2a2-531f-46ef-8d73-9475dbc47494 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.941350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-3183190e-3202-40be-894c-e47b186708b5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.941530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-3183190e-3202-40be-894c-e47b186708b5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.941687] env[61991]: DEBUG nova.network.neutron [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.944394] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e67ac19-f7df-4849-bdec-3d404bea3daf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.953060] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d652bc-bbdb-4978-9597-82ffc0413702 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.974116] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129841, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.974658] env[61991]: DEBUG nova.compute.provider_tree [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.002359] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526399a2-461c-3740-a00d-54cc781a1692, 'name': SearchDatastore_Task, 'duration_secs': 0.00959} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.002660] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.002913] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.003203] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.168483] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5246de71-1f78-783f-8129-65018f51f896, 'name': SearchDatastore_Task, 'duration_secs': 0.009034} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.169101] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1c35cbf-18bd-4e1b-887a-ceea2f1bb878 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.174323] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 848.174323] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5254868d-1116-33ec-0280-84ed788f9fed" [ 848.174323] env[61991]: _type = "Task" [ 848.174323] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.182039] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5254868d-1116-33ec-0280-84ed788f9fed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.347282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.350346] env[61991]: INFO nova.compute.manager [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Took 31.20 seconds to build instance. [ 848.400153] env[61991]: DEBUG oslo_vmware.api [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332482} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.400428] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.400616] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.400798] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.400970] env[61991]: INFO nova.compute.manager [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 848.401231] env[61991]: DEBUG oslo.service.loopingcall [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.401422] env[61991]: DEBUG nova.compute.manager [-] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.401516] env[61991]: DEBUG nova.network.neutron [-] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 848.467254] env[61991]: DEBUG oslo_vmware.api [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129841, 'name': PowerOnVM_Task, 'duration_secs': 0.875272} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.467600] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.467814] env[61991]: INFO nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Took 7.97 seconds to spawn the instance on the hypervisor. [ 848.467997] env[61991]: DEBUG nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.468787] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17267f4-43a7-4b89-8166-ab3d7e94105a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.479077] env[61991]: DEBUG nova.scheduler.client.report [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.498052] env[61991]: DEBUG nova.network.neutron [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 848.684563] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5254868d-1116-33ec-0280-84ed788f9fed, 'name': SearchDatastore_Task, 'duration_secs': 0.009264} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.685242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.685433] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. {{(pid=61991) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 848.685644] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.685904] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.686352] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cdb0714-3f86-40dc-bcce-521085b4c505 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.688553] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5a5bff7-edb6-4546-bd26-092bf3b8cd5b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.696485] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 848.696485] env[61991]: value = "task-1129845" [ 848.696485] env[61991]: _type = "Task" [ 848.696485] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.701976] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.702253] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.703782] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb48ab83-a216-40a8-a829-f474626f209e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.709790] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.712833] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 848.712833] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c8330d-8d73-159d-6a4d-a52d2b5ec0e9" [ 848.712833] env[61991]: _type = "Task" [ 848.712833] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.720959] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c8330d-8d73-159d-6a4d-a52d2b5ec0e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.853272] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8d0622f4-1d2e-4763-8e75-602a53d69246 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.713s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.909463] env[61991]: DEBUG nova.network.neutron [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Updating instance_info_cache with network_info: [{"id": "58dc3a16-cc46-440b-b310-026854eda3a2", "address": "fa:16:3e:b4:50:75", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58dc3a16-cc", "ovs_interfaceid": "58dc3a16-cc46-440b-b310-026854eda3a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.988884] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.989469] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.994331] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.794s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.996260] env[61991]: INFO nova.compute.claims [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.999540] env[61991]: INFO nova.compute.manager [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Took 30.60 seconds to build instance. [ 849.206621] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129845, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474335} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.206919] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. [ 849.207731] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3d01eb-8c60-4cf2-ad31-91b1443ce477 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.233347] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.236706] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed9e6055-81f2-4627-95a6-090e51c74138 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.255641] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c8330d-8d73-159d-6a4d-a52d2b5ec0e9, 'name': SearchDatastore_Task, 'duration_secs': 0.007944} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.257516] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 849.257516] env[61991]: value = "task-1129846" [ 849.257516] env[61991]: _type = "Task" [ 849.257516] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.257810] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad55058f-0cea-4fbb-8191-5c8276edd51a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.266149] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 849.266149] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52afb8d8-e454-7c89-bb85-da4ccbf8d1fc" [ 849.266149] env[61991]: _type = "Task" [ 849.266149] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.269484] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.283669] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52afb8d8-e454-7c89-bb85-da4ccbf8d1fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.412629] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-3183190e-3202-40be-894c-e47b186708b5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.414066] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance network_info: |[{"id": "58dc3a16-cc46-440b-b310-026854eda3a2", "address": "fa:16:3e:b4:50:75", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58dc3a16-cc", "ovs_interfaceid": "58dc3a16-cc46-440b-b310-026854eda3a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.414578] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:50:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58dc3a16-cc46-440b-b310-026854eda3a2', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.424338] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating folder: Project (2dddb51029854105bc6c9c8724181d39). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.425200] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09588c30-22fa-448f-828a-3364afa325f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.436887] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created folder: Project (2dddb51029854105bc6c9c8724181d39) in parent group-v246753. [ 849.437140] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating folder: Instances. Parent ref: group-v246894. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.437505] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9386f1f3-441b-474c-bc08-7fcd0c8b1bb6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.446632] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created folder: Instances in parent group-v246894. [ 849.446956] env[61991]: DEBUG oslo.service.loopingcall [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.447261] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.447485] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb821b0d-4196-4353-83af-3ee5f1713d98 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.471407] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.471407] env[61991]: value = "task-1129849" [ 849.471407] env[61991]: _type = "Task" [ 849.471407] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.479190] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129849, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.504778] env[61991]: DEBUG nova.compute.utils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.512195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec07d6c2-d8b2-4baa-bd9e-a65b037416a2 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.123s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.514678] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 849.514678] env[61991]: DEBUG nova.network.neutron [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.588119] env[61991]: DEBUG nova.compute.manager [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Received event network-changed-58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.588929] env[61991]: DEBUG nova.compute.manager [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Refreshing instance network info cache due to event network-changed-58dc3a16-cc46-440b-b310-026854eda3a2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.589392] env[61991]: DEBUG oslo_concurrency.lockutils [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] Acquiring lock "refresh_cache-3183190e-3202-40be-894c-e47b186708b5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.589673] env[61991]: DEBUG oslo_concurrency.lockutils [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] Acquired lock "refresh_cache-3183190e-3202-40be-894c-e47b186708b5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.590021] env[61991]: DEBUG nova.network.neutron [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Refreshing network info cache for port 58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 849.593540] env[61991]: DEBUG nova.policy [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e647f3055e254e2d9b48f976e710348c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c89707d8b26430c830449ab9bca4a62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.770517] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.781423] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52afb8d8-e454-7c89-bb85-da4ccbf8d1fc, 'name': SearchDatastore_Task, 'duration_secs': 0.020625} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.781423] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.781423] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.781423] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9bbfbbe2-5d17-42ce-83c0-e78c0f42a8c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.786895] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 849.786895] env[61991]: value = "task-1129850" [ 849.786895] env[61991]: _type = "Task" [ 849.786895] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.794967] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129850, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.852822] env[61991]: DEBUG nova.network.neutron [-] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.987143] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129849, 'name': CreateVM_Task, 'duration_secs': 0.496026} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.988595] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.991684] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.991864] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.993067] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.993310] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "f902f00f-eb2f-418a-bbfa-535442446681" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.993527] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "f902f00f-eb2f-418a-bbfa-535442446681" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.994952] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b26835a1-4d5d-4c5a-b7de-20e685fef3cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.001995] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 850.001995] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b29e5d-671b-adf4-abe2-b56f2a00099e" [ 850.001995] env[61991]: _type = "Task" [ 850.001995] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.013817] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b29e5d-671b-adf4-abe2-b56f2a00099e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.017344] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.275191] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129846, 'name': ReconfigVM_Task, 'duration_secs': 0.625569} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.275191] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Reconfigured VM instance instance-00000049 to attach disk [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.280380] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc37293-c827-4b2a-9b4e-82665ee2a034 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.315451] env[61991]: DEBUG nova.network.neutron [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Successfully created port: db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.319550] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-012b0ccd-b9d7-4c23-992a-bd218cea91aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.337105] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129850, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49192} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.338404] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 850.338939] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.339287] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 850.339287] env[61991]: value = "task-1129851" [ 850.339287] env[61991]: _type = "Task" [ 850.339287] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.339478] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac91ed83-ec6e-4687-9a10-8d642ff2c7c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.351471] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129851, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.352708] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 850.352708] env[61991]: value = "task-1129852" [ 850.352708] env[61991]: _type = "Task" [ 850.352708] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.355975] env[61991]: INFO nova.compute.manager [-] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Took 1.95 seconds to deallocate network for instance. [ 850.362251] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.440500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6869dc-6512-4db4-a544-6589dc0cea24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.448936] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fa64b2-863e-460b-9944-1ea44741e8d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.479330] env[61991]: DEBUG nova.network.neutron [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Updated VIF entry in instance network info cache for port 58dc3a16-cc46-440b-b310-026854eda3a2. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 850.479699] env[61991]: DEBUG nova.network.neutron [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Updating instance_info_cache with network_info: [{"id": "58dc3a16-cc46-440b-b310-026854eda3a2", "address": "fa:16:3e:b4:50:75", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58dc3a16-cc", "ovs_interfaceid": "58dc3a16-cc46-440b-b310-026854eda3a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.481860] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f1433e-e055-4add-bf25-b02cb0e8ece3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.490025] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edff425-2569-48cc-b9a9-57bae560171f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.506164] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.509169] env[61991]: DEBUG nova.compute.provider_tree [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.523275] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b29e5d-671b-adf4-abe2-b56f2a00099e, 'name': SearchDatastore_Task, 'duration_secs': 0.047516} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.523696] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.523940] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.524191] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.524486] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.524618] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.524983] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0db64f1e-d9dd-466a-9469-3b9b0cda275a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.543910] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.544139] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.544902] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e445e79b-9ac6-431d-83f2-2fd578e0e764 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.553462] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 850.553462] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525901ce-9ba3-a2ba-d692-052809aea9a6" [ 850.553462] env[61991]: _type = "Task" [ 850.553462] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.559828] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525901ce-9ba3-a2ba-d692-052809aea9a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.776186] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.776457] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.850795] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.860847] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.869960] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.985989] env[61991]: DEBUG oslo_concurrency.lockutils [req-71caaa0b-6e75-474d-a43a-3b60ed92e94b req-d2bdf598-c508-41f4-9279-540623a764ff service nova] Releasing lock "refresh_cache-3183190e-3202-40be-894c-e47b186708b5" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.032871] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.036437] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.040437] env[61991]: ERROR nova.scheduler.client.report [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [req-ef26aa58-b396-4277-b761-e5d81cfac50a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ef26aa58-b396-4277-b761-e5d81cfac50a"}]} [ 851.061688] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525901ce-9ba3-a2ba-d692-052809aea9a6, 'name': SearchDatastore_Task, 'duration_secs': 0.022657} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.062666] env[61991]: DEBUG nova.scheduler.client.report [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 851.067028] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.067267] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.067451] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.067654] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.067806] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.067958] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.068215] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.068343] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.068517] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.068675] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.068855] env[61991]: DEBUG nova.virt.hardware [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.070589] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0ab05d-5704-4c3f-be56-f129a9387e7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.074249] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc47935e-e20a-412d-87c9-bc507ced3e08 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.082225] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 851.082225] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a9cbe6-868f-479d-6a8f-d1824745ee65" [ 851.082225] env[61991]: _type = "Task" [ 851.082225] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.088381] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65aa4779-6b76-4a55-9f5c-3a82e8e6b15c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.092848] env[61991]: DEBUG nova.scheduler.client.report [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 851.093104] env[61991]: DEBUG nova.compute.provider_tree [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.106904] env[61991]: DEBUG nova.scheduler.client.report [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 851.112198] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a9cbe6-868f-479d-6a8f-d1824745ee65, 'name': SearchDatastore_Task, 'duration_secs': 0.00884} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.112950] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.113480] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.113940] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3edba905-34c4-41e5-8c3d-d3bd7c70f2db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.120510] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 851.120510] env[61991]: value = "task-1129853" [ 851.120510] env[61991]: _type = "Task" [ 851.120510] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.129438] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.130390] env[61991]: DEBUG nova.scheduler.client.report [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 851.280420] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.354284] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129851, 'name': ReconfigVM_Task, 'duration_secs': 0.68238} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.359958] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.360812] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-414f35a6-75b8-4791-9720-422050e7fce4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.367724] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129852, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.757174} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.369205] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.369589] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 851.369589] env[61991]: value = "task-1129854" [ 851.369589] env[61991]: _type = "Task" [ 851.369589] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.370317] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4690ba04-39a1-431a-bb7a-21123cedc8cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.385749] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129854, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.401894] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.404994] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a43aa9c0-0440-450d-8ab7-d71763b6e128 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.426071] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 851.426071] env[61991]: value = "task-1129855" [ 851.426071] env[61991]: _type = "Task" [ 851.426071] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.437610] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129855, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.451782] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85998da3-951d-41a1-a42d-a17dfc3068d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.463595] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d4e9d6-b60f-4209-afbb-cb654f3657f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.506074] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e412c033-5182-41fb-828c-2251812293fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.514753] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad055c4-7a05-452a-9571-5d04ea92452d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.532584] env[61991]: DEBUG nova.compute.manager [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.533049] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.533348] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.533540] env[61991]: DEBUG nova.compute.manager [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.534143] env[61991]: DEBUG nova.compute.provider_tree [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.536213] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089098bd-bea6-4f09-a37f-565aa6e80039 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.540885] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7b95e4-a720-4130-80fb-171382d102df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.552829] env[61991]: DEBUG nova.compute.manager [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 851.553569] env[61991]: DEBUG nova.objects.instance [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'flavor' on Instance uuid d36bd981-9ea3-46f6-8376-ac1e0c3bf61e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.611212] env[61991]: DEBUG nova.compute.manager [req-be8b2578-5513-46d9-b3c7-74c67736364d req-a10edee1-248a-45da-94af-624f04f43e66 service nova] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Received event network-vif-deleted-f9418974-50f5-48bc-a8cc-dd016b21728a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.630410] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451103} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.630705] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.630947] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.631238] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72b5cfb6-c5af-4ea7-af30-3a1efc7df392 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.637315] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 851.637315] env[61991]: value = "task-1129856" [ 851.637315] env[61991]: _type = "Task" [ 851.637315] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.645942] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.804489] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.884038] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129854, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.935200] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129855, 'name': ReconfigVM_Task, 'duration_secs': 0.442026} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.935507] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03/96627657-d3d9-4a64-b83e-87323bd10c03.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.936122] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52c24901-816a-45b3-ad61-bb5aac582688 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.942142] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 851.942142] env[61991]: value = "task-1129857" [ 851.942142] env[61991]: _type = "Task" [ 851.942142] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.950265] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129857, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.976840] env[61991]: DEBUG nova.network.neutron [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Successfully updated port: db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.056615] env[61991]: INFO nova.compute.manager [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] instance snapshotting [ 852.061641] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.062121] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54151a33-88c0-4942-a46c-ecafec555f10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.064443] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a7961f-fa34-4672-950d-d45d943dac84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.085500] env[61991]: DEBUG nova.scheduler.client.report [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 852.085776] env[61991]: DEBUG nova.compute.provider_tree [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 103 to 104 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 852.085962] env[61991]: DEBUG nova.compute.provider_tree [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 852.089929] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb36496-c523-4a16-b29a-8843be367ff8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.092909] env[61991]: DEBUG oslo_vmware.api [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 852.092909] env[61991]: value = "task-1129858" [ 852.092909] env[61991]: _type = "Task" [ 852.092909] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.105739] env[61991]: DEBUG oslo_vmware.api [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.153022] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087647} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.153022] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.153022] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcedd0ce-30e3-467d-a0f2-ab0c490f9e7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.176550] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.177263] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fc197e3-6829-4161-bb60-e04a3e16ff68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.196957] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 852.196957] env[61991]: value = "task-1129859" [ 852.196957] env[61991]: _type = "Task" [ 852.196957] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.206520] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.382759] env[61991]: DEBUG oslo_vmware.api [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129854, 'name': PowerOnVM_Task, 'duration_secs': 0.536517} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.383690] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.386254] env[61991]: DEBUG nova.compute.manager [None req-b3d9ff06-0f4d-4f47-b6b7-0e86b4e90768 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.387023] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d8246d-15ee-45fc-b9d4-29609b4cc6b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.451567] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129857, 'name': Rename_Task, 'duration_secs': 0.240109} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.451849] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.452259] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83dacd25-59e1-4072-9a6c-56e55e3d469e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.457975] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 852.457975] env[61991]: value = "task-1129860" [ 852.457975] env[61991]: _type = "Task" [ 852.457975] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.465225] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.479923] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.480049] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.480186] env[61991]: DEBUG nova.network.neutron [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.594739] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.600s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.595359] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.598492] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.436s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.599884] env[61991]: INFO nova.compute.claims [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.606848] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 852.606848] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f97d4865-a6ad-4f7e-ba05-91eb0c4b5972 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.616195] env[61991]: DEBUG oslo_vmware.api [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129858, 'name': PowerOffVM_Task, 'duration_secs': 0.33049} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.617087] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.617214] env[61991]: DEBUG nova.compute.manager [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.617648] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 852.617648] env[61991]: value = "task-1129861" [ 852.617648] env[61991]: _type = "Task" [ 852.617648] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.618279] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2707048f-1be8-4db6-8a63-1304b504e0af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.632415] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129861, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.708956] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.967762] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129860, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.033426] env[61991]: DEBUG nova.network.neutron [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.109099] env[61991]: DEBUG nova.compute.utils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.110650] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.110823] env[61991]: DEBUG nova.network.neutron [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 853.130708] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129861, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.135175] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6de54adc-df36-4e66-9a46-6854c25bcfc1 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.602s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.203506] env[61991]: DEBUG nova.policy [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e647f3055e254e2d9b48f976e710348c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c89707d8b26430c830449ab9bca4a62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.210190] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129859, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.274696] env[61991]: DEBUG nova.network.neutron [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updating instance_info_cache with network_info: [{"id": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "address": "fa:16:3e:14:21:d4", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6f94dd-10", "ovs_interfaceid": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.469140] env[61991]: DEBUG oslo_vmware.api [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129860, 'name': PowerOnVM_Task, 'duration_secs': 0.946505} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.470069] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.470361] env[61991]: DEBUG nova.compute.manager [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.472263] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c43382-a238-49c2-bbb6-aea8d4adc579 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.555040] env[61991]: DEBUG nova.network.neutron [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Successfully created port: ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.614911] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.633060] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129861, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.658140] env[61991]: DEBUG nova.compute.manager [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Received event network-vif-plugged-db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.658140] env[61991]: DEBUG oslo_concurrency.lockutils [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] Acquiring lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.658140] env[61991]: DEBUG oslo_concurrency.lockutils [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.658140] env[61991]: DEBUG oslo_concurrency.lockutils [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.658696] env[61991]: DEBUG nova.compute.manager [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] No waiting events found dispatching network-vif-plugged-db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.659129] env[61991]: WARNING nova.compute.manager [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Received unexpected event network-vif-plugged-db6f94dd-1023-4cf6-8997-5bead8bfa1b2 for instance with vm_state building and task_state spawning. [ 853.659590] env[61991]: DEBUG nova.compute.manager [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Received event network-changed-db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.661363] env[61991]: DEBUG nova.compute.manager [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Refreshing instance network info cache due to event network-changed-db6f94dd-1023-4cf6-8997-5bead8bfa1b2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 853.661363] env[61991]: DEBUG oslo_concurrency.lockutils [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] Acquiring lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.708289] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129859, 'name': ReconfigVM_Task, 'duration_secs': 1.027897} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.711019] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.711841] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fb82693-080e-4afa-89a1-8213a6f0ebda {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.718979] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 853.718979] env[61991]: value = "task-1129862" [ 853.718979] env[61991]: _type = "Task" [ 853.718979] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.730198] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129862, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.763057] env[61991]: INFO nova.compute.manager [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Unrescuing [ 853.763148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.763897] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquired lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.763897] env[61991]: DEBUG nova.network.neutron [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.777485] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.778340] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Instance network_info: |[{"id": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "address": "fa:16:3e:14:21:d4", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6f94dd-10", "ovs_interfaceid": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.778561] env[61991]: DEBUG oslo_concurrency.lockutils [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] Acquired lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.779813] env[61991]: DEBUG nova.network.neutron [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Refreshing network info cache for port db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.781457] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:21:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db6f94dd-1023-4cf6-8997-5bead8bfa1b2', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.789782] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating folder: Project (8c89707d8b26430c830449ab9bca4a62). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.793503] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-016dc93f-1e15-4973-aba7-8553ea3fae96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.804329] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created folder: Project (8c89707d8b26430c830449ab9bca4a62) in parent group-v246753. [ 853.804529] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating folder: Instances. Parent ref: group-v246897. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 853.804768] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fc02ec3-88e6-4318-8d6a-e2ce64c4db47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.814636] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created folder: Instances in parent group-v246897. [ 853.814725] env[61991]: DEBUG oslo.service.loopingcall [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.815624] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.815624] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7db87b1d-d4fe-4cdc-a0fe-05118f3dc34c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.838832] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.838832] env[61991]: value = "task-1129865" [ 853.838832] env[61991]: _type = "Task" [ 853.838832] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.846739] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129865, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.987767] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.989583] env[61991]: DEBUG nova.objects.instance [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'flavor' on Instance uuid d36bd981-9ea3-46f6-8376-ac1e0c3bf61e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.995594] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0370dc-d4e8-4b6a-8b40-bdda70a1b4dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.004946] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ae8965-d4f2-4774-a767-220e58e2d3a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.038204] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18965ab-b930-49da-b12d-2261b5951a0c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.048772] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847289c0-5b79-49e2-8eec-83dbe87dbfac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.062805] env[61991]: DEBUG nova.compute.provider_tree [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.135878] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129861, 'name': CreateSnapshot_Task, 'duration_secs': 1.345438} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.136655] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 854.137528] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e898a4-edc4-487b-a866-47b592ac0c1a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.229021] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129862, 'name': Rename_Task, 'duration_secs': 0.156737} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.229415] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.229666] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6084f4a8-71a9-4439-bf4d-de2958eb41b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.236438] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 854.236438] env[61991]: value = "task-1129866" [ 854.236438] env[61991]: _type = "Task" [ 854.236438] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.244071] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129866, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.259230] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "96627657-d3d9-4a64-b83e-87323bd10c03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.259697] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "96627657-d3d9-4a64-b83e-87323bd10c03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.260122] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "96627657-d3d9-4a64-b83e-87323bd10c03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.260369] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "96627657-d3d9-4a64-b83e-87323bd10c03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.260614] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "96627657-d3d9-4a64-b83e-87323bd10c03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.262851] env[61991]: INFO nova.compute.manager [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Terminating instance [ 854.264806] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "refresh_cache-96627657-d3d9-4a64-b83e-87323bd10c03" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.265051] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquired lock "refresh_cache-96627657-d3d9-4a64-b83e-87323bd10c03" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.265414] env[61991]: DEBUG nova.network.neutron [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.349660] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129865, 'name': CreateVM_Task, 'duration_secs': 0.359441} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.349830] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.350581] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.350780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.351114] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.351379] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c39093-b56b-4296-be31-5547c2e2dec4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.356542] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 854.356542] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5208bdec-3937-dc17-341f-524eb8fffc73" [ 854.356542] env[61991]: _type = "Task" [ 854.356542] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.369390] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5208bdec-3937-dc17-341f-524eb8fffc73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.475197] env[61991]: DEBUG nova.network.neutron [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updating instance_info_cache with network_info: [{"id": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "address": "fa:16:3e:81:55:e0", "network": {"id": "3005bb96-30a5-42cb-9cbc-05a63bc73c89", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-382153796-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "79c013b6c254403b91f520e2af18027f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0df968ae-c1ef-4009-a0f4-6f2e799c2fda", "external-id": "nsx-vlan-transportzone-864", "segmentation_id": 864, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87e2f32e-53", "ovs_interfaceid": "87e2f32e-53df-49a9-bae3-f9529f2d3c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.494506] env[61991]: DEBUG oslo_concurrency.lockutils [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.494739] env[61991]: DEBUG oslo_concurrency.lockutils [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.495015] env[61991]: DEBUG nova.network.neutron [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.495281] env[61991]: DEBUG nova.objects.instance [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'info_cache' on Instance uuid d36bd981-9ea3-46f6-8376-ac1e0c3bf61e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.566275] env[61991]: DEBUG nova.network.neutron [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updated VIF entry in instance network info cache for port db6f94dd-1023-4cf6-8997-5bead8bfa1b2. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 854.566787] env[61991]: DEBUG nova.network.neutron [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updating instance_info_cache with network_info: [{"id": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "address": "fa:16:3e:14:21:d4", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6f94dd-10", "ovs_interfaceid": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.568638] env[61991]: DEBUG nova.scheduler.client.report [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.630378] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.658804] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 854.661115] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c3d03d65-2c6c-4ea7-9c82-98f207abd33d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.670099] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 854.670099] env[61991]: value = "task-1129867" [ 854.670099] env[61991]: _type = "Task" [ 854.670099] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.672353] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.672796] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.673093] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.673414] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.674020] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.674020] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.674147] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.674234] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.674405] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.674634] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.674745] env[61991]: DEBUG nova.virt.hardware [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.675693] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1662be1-0902-4f5a-8672-0ad75b29bdc1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.686300] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129867, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.689358] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d449f9-fb11-499a-849d-db8e33e9e6ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.746434] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129866, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.784420] env[61991]: DEBUG nova.network.neutron [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.842980] env[61991]: DEBUG nova.network.neutron [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.871999] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5208bdec-3937-dc17-341f-524eb8fffc73, 'name': SearchDatastore_Task, 'duration_secs': 0.014627} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.871999] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.871999] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.872327] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.872429] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.872630] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.873205] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cbea6b4-1161-4e62-8afb-bdb8da29489b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.884051] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.884051] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.884402] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3918241e-c0f5-4d7c-ae20-92f68cde5cea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.896929] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 854.896929] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520b477e-11bd-b279-9080-8e813a7860c5" [ 854.896929] env[61991]: _type = "Task" [ 854.896929] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.906505] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520b477e-11bd-b279-9080-8e813a7860c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.981185] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Releasing lock "refresh_cache-c1d4d2eb-810c-42f9-a244-e5aed6e8d517" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.981972] env[61991]: DEBUG nova.objects.instance [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lazy-loading 'flavor' on Instance uuid c1d4d2eb-810c-42f9-a244-e5aed6e8d517 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.998503] env[61991]: DEBUG nova.objects.base [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 855.073958] env[61991]: DEBUG oslo_concurrency.lockutils [req-77f7bd2b-3e2b-418a-b8b1-2ef608e1cede req-ac42b04f-4a09-425e-a2c0-29f10be2ec70 service nova] Releasing lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.078259] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.079149] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.084571] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.418s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.086878] env[61991]: INFO nova.compute.claims [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.188889] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129867, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.247401] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129866, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.302304] env[61991]: DEBUG nova.network.neutron [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Successfully updated port: ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.345726] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Releasing lock "refresh_cache-96627657-d3d9-4a64-b83e-87323bd10c03" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.346157] env[61991]: DEBUG nova.compute.manager [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 855.346369] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.347295] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b818eb1b-d356-47b3-b730-672d1534c676 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.354825] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.355108] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8e3a0b9-fc48-40bc-a72d-4e895d46abab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.361432] env[61991]: DEBUG oslo_vmware.api [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 855.361432] env[61991]: value = "task-1129868" [ 855.361432] env[61991]: _type = "Task" [ 855.361432] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.370810] env[61991]: DEBUG oslo_vmware.api [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.408045] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520b477e-11bd-b279-9080-8e813a7860c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009851} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.408871] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-966d44a2-a2d0-4c5c-ad8a-ce8de03f09a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.414598] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 855.414598] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5223c747-90af-d716-b19c-3f08bde24cff" [ 855.414598] env[61991]: _type = "Task" [ 855.414598] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.422599] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5223c747-90af-d716-b19c-3f08bde24cff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.488427] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e830ec8c-1e66-4219-a023-0aba6867acb9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.511980] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.512422] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4debd28-bf85-4dc1-8fd7-c294dc187ee1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.519873] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 855.519873] env[61991]: value = "task-1129869" [ 855.519873] env[61991]: _type = "Task" [ 855.519873] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.528691] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.594656] env[61991]: DEBUG nova.compute.utils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.597141] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.597141] env[61991]: DEBUG nova.network.neutron [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.687442] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129867, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.697746] env[61991]: DEBUG nova.policy [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3005c87b7d245cd8a35fdab04e808e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '537c0b730d2d47498fcf8a8ddd238c3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.749147] env[61991]: DEBUG oslo_vmware.api [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129866, 'name': PowerOnVM_Task, 'duration_secs': 1.026931} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.749426] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.749630] env[61991]: INFO nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Took 9.08 seconds to spawn the instance on the hypervisor. [ 855.749811] env[61991]: DEBUG nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.750594] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917f185d-9dc7-47b5-b7b5-2efd4326cb55 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.794080] env[61991]: DEBUG nova.compute.manager [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Received event network-vif-plugged-ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.794080] env[61991]: DEBUG oslo_concurrency.lockutils [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] Acquiring lock "f7eab1a9-2a8a-4664-8736-ca599e207c31-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.794080] env[61991]: DEBUG oslo_concurrency.lockutils [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.794080] env[61991]: DEBUG oslo_concurrency.lockutils [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.794080] env[61991]: DEBUG nova.compute.manager [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] No waiting events found dispatching network-vif-plugged-ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.794080] env[61991]: WARNING nova.compute.manager [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Received unexpected event network-vif-plugged-ce8085ee-47de-4c70-aafc-dcdf66c17383 for instance with vm_state building and task_state spawning. [ 855.794080] env[61991]: DEBUG nova.compute.manager [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Received event network-changed-ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.794080] env[61991]: DEBUG nova.compute.manager [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Refreshing instance network info cache due to event network-changed-ce8085ee-47de-4c70-aafc-dcdf66c17383. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.794412] env[61991]: DEBUG oslo_concurrency.lockutils [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] Acquiring lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.794412] env[61991]: DEBUG oslo_concurrency.lockutils [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] Acquired lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.794412] env[61991]: DEBUG nova.network.neutron [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Refreshing network info cache for port ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 855.801542] env[61991]: DEBUG nova.network.neutron [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.805037] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.872706] env[61991]: DEBUG oslo_vmware.api [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129868, 'name': PowerOffVM_Task, 'duration_secs': 0.214425} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.872995] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 855.873238] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 855.873503] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54d3dbb7-9a9a-4380-af89-6698fef90c4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.903796] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 855.904093] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 855.904288] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Deleting the datastore file [datastore1] 96627657-d3d9-4a64-b83e-87323bd10c03 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.904617] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6cd016b-0cbd-4fcc-b4cf-c921082dd781 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.912015] env[61991]: DEBUG oslo_vmware.api [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for the task: (returnval){ [ 855.912015] env[61991]: value = "task-1129871" [ 855.912015] env[61991]: _type = "Task" [ 855.912015] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.931964] env[61991]: DEBUG oslo_vmware.api [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.935489] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5223c747-90af-d716-b19c-3f08bde24cff, 'name': SearchDatastore_Task, 'duration_secs': 0.010553} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.935727] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.935989] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/56e0ec0a-61d5-4394-a8f4-1a52ab61e08b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.936299] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-197b286a-385a-4d6d-85e9-dc7c9b89941e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.942768] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 855.942768] env[61991]: value = "task-1129872" [ 855.942768] env[61991]: _type = "Task" [ 855.942768] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.950495] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.030049] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129869, 'name': PowerOffVM_Task, 'duration_secs': 0.205153} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.030364] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 856.036254] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 856.036667] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59f7188e-3fb5-4ac9-808f-e632d8b840b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.058598] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 856.058598] env[61991]: value = "task-1129873" [ 856.058598] env[61991]: _type = "Task" [ 856.058598] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.070255] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129873, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.100083] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.186919] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129867, 'name': CloneVM_Task, 'duration_secs': 1.259225} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.189860] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Created linked-clone VM from snapshot [ 856.190984] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00f8791-80f9-4438-b5a5-2cece83a0b6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.199908] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Uploading image 792a07de-7014-4286-a4ae-8dd9b77c3d34 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 856.223526] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 856.223852] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dfd8d121-8c80-4774-8f5e-80ed2707b2a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.230572] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 856.230572] env[61991]: value = "task-1129874" [ 856.230572] env[61991]: _type = "Task" [ 856.230572] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.242467] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129874, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.272524] env[61991]: INFO nova.compute.manager [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Took 34.42 seconds to build instance. [ 856.305195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.339355] env[61991]: DEBUG nova.network.neutron [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.352562] env[61991]: DEBUG nova.network.neutron [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Successfully created port: 490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.423166] env[61991]: DEBUG oslo_vmware.api [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Task: {'id': task-1129871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096689} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.423489] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.423946] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.423946] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.424123] env[61991]: INFO nova.compute.manager [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Took 1.08 seconds to destroy the instance on the hypervisor. [ 856.424295] env[61991]: DEBUG oslo.service.loopingcall [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.424515] env[61991]: DEBUG nova.compute.manager [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.424655] env[61991]: DEBUG nova.network.neutron [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 856.450542] env[61991]: DEBUG nova.network.neutron [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.459298] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129872, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.466124] env[61991]: DEBUG nova.network.neutron [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.473342] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6c46b5-2ff0-4182-8727-5b9e5c596e23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.481422] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb09420-cf14-480a-929b-c60d7182e956 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.514515] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b6570a-d1a5-454c-9458-d5840b671cf9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.522287] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df9b25a-aaa0-473d-a147-d9ef674b9a06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.537261] env[61991]: DEBUG nova.compute.provider_tree [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 856.568320] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129873, 'name': ReconfigVM_Task, 'duration_secs': 0.34546} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.568624] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 856.568792] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.569045] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0950941a-7611-4c62-8967-a2e24f9d55c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.575369] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 856.575369] env[61991]: value = "task-1129875" [ 856.575369] env[61991]: _type = "Task" [ 856.575369] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.583833] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129875, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.741708] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129874, 'name': Destroy_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.776117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5197ca35-7c95-4144-9149-b054fc6729a9 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "3183190e-3202-40be-894c-e47b186708b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.940s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.809470] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.809803] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8870921-c553-4aeb-8b4b-7c85c14dc976 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.817456] env[61991]: DEBUG oslo_vmware.api [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 856.817456] env[61991]: value = "task-1129876" [ 856.817456] env[61991]: _type = "Task" [ 856.817456] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.825927] env[61991]: DEBUG oslo_vmware.api [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.955279] env[61991]: DEBUG nova.network.neutron [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.957273] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526333} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.957399] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/56e0ec0a-61d5-4394-a8f4-1a52ab61e08b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.957542] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.957757] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07f407b2-6e8d-4439-883d-d5c4518aa7dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.964957] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 856.964957] env[61991]: value = "task-1129877" [ 856.964957] env[61991]: _type = "Task" [ 856.964957] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.970183] env[61991]: DEBUG oslo_concurrency.lockutils [req-804ec7a7-987e-48cb-ac2d-5523e449f4a9 req-f3693eaa-9c44-415d-ac39-a8f87c69c3bc service nova] Releasing lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.970183] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.970183] env[61991]: DEBUG nova.network.neutron [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.976451] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129877, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.076051] env[61991]: DEBUG nova.scheduler.client.report [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 104 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 857.076676] env[61991]: DEBUG nova.compute.provider_tree [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 104 to 105 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 857.076676] env[61991]: DEBUG nova.compute.provider_tree [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 857.088557] env[61991]: DEBUG oslo_vmware.api [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129875, 'name': PowerOnVM_Task, 'duration_secs': 0.462211} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.088830] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.089084] env[61991]: DEBUG nova.compute.manager [None req-c0242524-f6e4-4eec-b376-1161e95da048 tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.089849] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e3c5c7-6b0b-4cf3-8dec-87e755d9822a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.110698] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.132880] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.133240] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.133422] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.133615] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.133763] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.133911] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.134132] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.134305] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.134470] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.134680] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.134863] env[61991]: DEBUG nova.virt.hardware [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.136207] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89248e1-bbc3-45fa-9e3b-3469ff4c5187 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.145581] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5524319-fe37-4a9d-a25a-116a5c418c20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.242050] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129874, 'name': Destroy_Task, 'duration_secs': 0.863581} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.242359] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Destroyed the VM [ 857.242600] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 857.242862] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c2d8cf39-8473-4ffd-96e5-c9132d39e722 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.249650] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 857.249650] env[61991]: value = "task-1129878" [ 857.249650] env[61991]: _type = "Task" [ 857.249650] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.258198] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129878, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.327684] env[61991]: DEBUG oslo_vmware.api [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129876, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.397115] env[61991]: INFO nova.compute.manager [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Rebuilding instance [ 857.438781] env[61991]: DEBUG nova.compute.manager [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.439669] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76be8e9-e87f-4306-863d-1bf4ea621b6c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.460958] env[61991]: INFO nova.compute.manager [-] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Took 1.04 seconds to deallocate network for instance. [ 857.476591] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129877, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.173698} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.476729] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.478027] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7470b91a-1d08-4564-bb6f-5d02ea8aaf53 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.500495] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/56e0ec0a-61d5-4394-a8f4-1a52ab61e08b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.501525] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4f26fde-045c-458f-be93-2a299629319f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.521558] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 857.521558] env[61991]: value = "task-1129879" [ 857.521558] env[61991]: _type = "Task" [ 857.521558] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.529977] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129879, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.534973] env[61991]: DEBUG nova.network.neutron [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.584561] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.585255] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.588148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.920s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.592471] env[61991]: INFO nova.compute.claims [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.719474] env[61991]: DEBUG nova.network.neutron [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Updating instance_info_cache with network_info: [{"id": "ce8085ee-47de-4c70-aafc-dcdf66c17383", "address": "fa:16:3e:3c:85:e5", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce8085ee-47", "ovs_interfaceid": "ce8085ee-47de-4c70-aafc-dcdf66c17383", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.759685] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129878, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.829684] env[61991]: DEBUG oslo_vmware.api [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129876, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.861038] env[61991]: DEBUG nova.compute.manager [req-865996ba-a327-488e-874c-de9ed71adf19 req-324f3f1b-2e02-4232-9986-a6067ca5c5f6 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Received event network-vif-plugged-490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.861292] env[61991]: DEBUG oslo_concurrency.lockutils [req-865996ba-a327-488e-874c-de9ed71adf19 req-324f3f1b-2e02-4232-9986-a6067ca5c5f6 service nova] Acquiring lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.861536] env[61991]: DEBUG oslo_concurrency.lockutils [req-865996ba-a327-488e-874c-de9ed71adf19 req-324f3f1b-2e02-4232-9986-a6067ca5c5f6 service nova] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.861716] env[61991]: DEBUG oslo_concurrency.lockutils [req-865996ba-a327-488e-874c-de9ed71adf19 req-324f3f1b-2e02-4232-9986-a6067ca5c5f6 service nova] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.861903] env[61991]: DEBUG nova.compute.manager [req-865996ba-a327-488e-874c-de9ed71adf19 req-324f3f1b-2e02-4232-9986-a6067ca5c5f6 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] No waiting events found dispatching network-vif-plugged-490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.862171] env[61991]: WARNING nova.compute.manager [req-865996ba-a327-488e-874c-de9ed71adf19 req-324f3f1b-2e02-4232-9986-a6067ca5c5f6 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Received unexpected event network-vif-plugged-490686f3-4e55-4f9f-afe0-6dbb260e6bea for instance with vm_state building and task_state spawning. [ 857.950178] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.950705] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-357797c3-a2f3-4a6b-97e1-0b096fc367c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.957707] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 857.957707] env[61991]: value = "task-1129880" [ 857.957707] env[61991]: _type = "Task" [ 857.957707] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.958015] env[61991]: DEBUG nova.network.neutron [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Successfully updated port: 490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.967306] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.970802] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.031023] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129879, 'name': ReconfigVM_Task, 'duration_secs': 0.416266} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.031920] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/56e0ec0a-61d5-4394-a8f4-1a52ab61e08b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.032583] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7824bf7-f00a-41d8-92b2-5abccba2e424 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.038820] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 858.038820] env[61991]: value = "task-1129881" [ 858.038820] env[61991]: _type = "Task" [ 858.038820] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.047860] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129881, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.097957] env[61991]: DEBUG nova.compute.utils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.099658] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.099857] env[61991]: DEBUG nova.network.neutron [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.156166] env[61991]: DEBUG nova.policy [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.222666] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.223232] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Instance network_info: |[{"id": "ce8085ee-47de-4c70-aafc-dcdf66c17383", "address": "fa:16:3e:3c:85:e5", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce8085ee-47", "ovs_interfaceid": "ce8085ee-47de-4c70-aafc-dcdf66c17383", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.223877] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:85:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce8085ee-47de-4c70-aafc-dcdf66c17383', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.233228] env[61991]: DEBUG oslo.service.loopingcall [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.233590] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.233899] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1bd2fe5-dde8-4c16-b69c-4b674a3b58a2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.264319] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129878, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.265836] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.265836] env[61991]: value = "task-1129882" [ 858.265836] env[61991]: _type = "Task" [ 858.265836] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.274094] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129882, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.327834] env[61991]: DEBUG oslo_vmware.api [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1129876, 'name': PowerOnVM_Task, 'duration_secs': 1.297307} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.328156] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.328621] env[61991]: DEBUG nova.compute.manager [None req-171bdf9d-673a-4fe2-a5f5-c3fdb2e32ba8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.329437] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1288cc9e-5148-4a97-8597-4ae9d1798a5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.467020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "refresh_cache-7634e00f-48ee-40a0-8625-2f8dc83cd58d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.467020] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired lock "refresh_cache-7634e00f-48ee-40a0-8625-2f8dc83cd58d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.467020] env[61991]: DEBUG nova.network.neutron [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.476799] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129880, 'name': PowerOffVM_Task, 'duration_secs': 0.240338} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.477346] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.477346] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.478117] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2573c1-db00-403e-bed8-baccb1133869 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.485711] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.486570] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8376f4aa-031d-495d-8a21-5601dcd30379 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.546222] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.546510] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.546689] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore1] 3183190e-3202-40be-894c-e47b186708b5 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.546953] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cb83fb0-b11d-42d0-adc7-fddec8cb82a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.552539] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129881, 'name': Rename_Task, 'duration_secs': 0.22005} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.553163] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.553408] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-880b9052-7634-454d-a4de-55a98d79aa01 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.556485] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 858.556485] env[61991]: value = "task-1129884" [ 858.556485] env[61991]: _type = "Task" [ 858.556485] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.561061] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 858.561061] env[61991]: value = "task-1129885" [ 858.561061] env[61991]: _type = "Task" [ 858.561061] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.568025] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.572394] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129885, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.606322] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.634965] env[61991]: DEBUG nova.network.neutron [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Successfully created port: f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.765006] env[61991]: DEBUG oslo_vmware.api [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129878, 'name': RemoveSnapshot_Task, 'duration_secs': 1.030564} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.768005] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 858.782554] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129882, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.930876] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.931149] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.931361] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.931568] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.931741] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.933652] env[61991]: INFO nova.compute.manager [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Terminating instance [ 858.935355] env[61991]: DEBUG nova.compute.manager [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 858.935557] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.936388] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ef4a50-50bb-46b7-abe5-af890da94883 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.945993] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.946564] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79de4627-dd22-4559-bed1-10c56f17a618 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.952934] env[61991]: DEBUG oslo_vmware.api [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 858.952934] env[61991]: value = "task-1129886" [ 858.952934] env[61991]: _type = "Task" [ 858.952934] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.954463] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb5ed84-1636-4d06-91d9-bdd9545b58d4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.966344] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bda551e-3e20-469c-ba2b-2d153680a7ae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.969427] env[61991]: DEBUG oslo_vmware.api [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129886, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.001132] env[61991]: DEBUG nova.network.neutron [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 859.003638] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a2a271-755f-4d29-a022-b0403d63084c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.011287] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78fd5a0-87fe-48ea-b7e6-c3a6cd1e13c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.024802] env[61991]: DEBUG nova.compute.provider_tree [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.069566] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413276} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.070348] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.070546] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 859.070727] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 859.076094] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129885, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.190535] env[61991]: DEBUG nova.network.neutron [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Updating instance_info_cache with network_info: [{"id": "490686f3-4e55-4f9f-afe0-6dbb260e6bea", "address": "fa:16:3e:cb:98:e2", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap490686f3-4e", "ovs_interfaceid": "490686f3-4e55-4f9f-afe0-6dbb260e6bea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.272968] env[61991]: WARNING nova.compute.manager [None req-ba0f8313-7389-400b-bc9a-1b3888a3cac5 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Image not found during snapshot: nova.exception.ImageNotFound: Image 792a07de-7014-4286-a4ae-8dd9b77c3d34 could not be found. [ 859.288967] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129882, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.465202] env[61991]: DEBUG oslo_vmware.api [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129886, 'name': PowerOffVM_Task, 'duration_secs': 0.419349} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.465532] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.466276] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.466276] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10f28ff4-7508-41b0-b334-da10ccf486a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.529715] env[61991]: DEBUG nova.scheduler.client.report [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.537925] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.538607] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.538938] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Deleting the datastore file [datastore1] c1d4d2eb-810c-42f9-a244-e5aed6e8d517 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.539379] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdd6aacb-56de-461d-9fd1-960825abad67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.552784] env[61991]: DEBUG oslo_vmware.api [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 859.552784] env[61991]: value = "task-1129888" [ 859.552784] env[61991]: _type = "Task" [ 859.552784] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.561800] env[61991]: DEBUG oslo_vmware.api [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.571814] env[61991]: DEBUG oslo_vmware.api [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129885, 'name': PowerOnVM_Task, 'duration_secs': 0.896304} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.571814] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.572927] env[61991]: INFO nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Took 8.54 seconds to spawn the instance on the hypervisor. [ 859.572927] env[61991]: DEBUG nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.573206] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7187ba87-43fd-4f20-a778-38dfdf5d3229 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.617242] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.643510] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.643762] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.643920] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.644156] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.644317] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.644520] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.644686] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.644854] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.645211] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.645495] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.645693] env[61991]: DEBUG nova.virt.hardware [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.646693] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a220809-e659-4036-968c-d9b32610b951 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.655848] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebaa25c5-720b-445f-b596-4eaf2cc619d3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.693572] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Releasing lock "refresh_cache-7634e00f-48ee-40a0-8625-2f8dc83cd58d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.693947] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Instance network_info: |[{"id": "490686f3-4e55-4f9f-afe0-6dbb260e6bea", "address": "fa:16:3e:cb:98:e2", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap490686f3-4e", "ovs_interfaceid": "490686f3-4e55-4f9f-afe0-6dbb260e6bea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.695748] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:98:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '490686f3-4e55-4f9f-afe0-6dbb260e6bea', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.703687] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating folder: Project (537c0b730d2d47498fcf8a8ddd238c3b). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.704065] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f5cb251-c397-435a-8721-aac5ec2fbdc7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.709071] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.709780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.709780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.709955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.710540] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.713011] env[61991]: INFO nova.compute.manager [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Terminating instance [ 859.715996] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Created folder: Project (537c0b730d2d47498fcf8a8ddd238c3b) in parent group-v246753. [ 859.716353] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating folder: Instances. Parent ref: group-v246903. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.716828] env[61991]: DEBUG nova.compute.manager [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.717038] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.717649] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0786cbb-4add-4c90-9a0c-8563c45d627f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.719796] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b3c013-8166-4d96-a343-406cad3d6acc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.728015] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.728283] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8538c020-6316-49e3-8ca7-020b8f38402e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.731421] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Created folder: Instances in parent group-v246903. [ 859.731677] env[61991]: DEBUG oslo.service.loopingcall [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.731870] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.732109] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e23f263-896c-409d-b676-630b9e0227d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.749833] env[61991]: DEBUG oslo_vmware.api [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 859.749833] env[61991]: value = "task-1129891" [ 859.749833] env[61991]: _type = "Task" [ 859.749833] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.755056] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.755056] env[61991]: value = "task-1129892" [ 859.755056] env[61991]: _type = "Task" [ 859.755056] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.762864] env[61991]: DEBUG oslo_vmware.api [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129891, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.768358] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129892, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.783683] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129882, 'name': CreateVM_Task, 'duration_secs': 1.116149} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.783875] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.784672] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.784849] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.785311] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.785702] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7d9e136-81e9-45cb-ab11-ef2c7c63ced9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.791100] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 859.791100] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a36e9-b0f0-5f22-87d1-275d396e7ffe" [ 859.791100] env[61991]: _type = "Task" [ 859.791100] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.802107] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a36e9-b0f0-5f22-87d1-275d396e7ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.947368] env[61991]: DEBUG nova.compute.manager [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Received event network-changed-490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.947600] env[61991]: DEBUG nova.compute.manager [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Refreshing instance network info cache due to event network-changed-490686f3-4e55-4f9f-afe0-6dbb260e6bea. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.947825] env[61991]: DEBUG oslo_concurrency.lockutils [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] Acquiring lock "refresh_cache-7634e00f-48ee-40a0-8625-2f8dc83cd58d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.947995] env[61991]: DEBUG oslo_concurrency.lockutils [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] Acquired lock "refresh_cache-7634e00f-48ee-40a0-8625-2f8dc83cd58d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.948197] env[61991]: DEBUG nova.network.neutron [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Refreshing network info cache for port 490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 860.034982] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.035588] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 860.040697] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.908s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.040936] env[61991]: DEBUG nova.objects.instance [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lazy-loading 'resources' on Instance uuid 15215bba-8160-408b-9e8f-afbb9e71d50e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.061105] env[61991]: DEBUG oslo_vmware.api [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254162} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.061429] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.061675] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.061911] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.062155] env[61991]: INFO nova.compute.manager [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Took 1.13 seconds to destroy the instance on the hypervisor. [ 860.062481] env[61991]: DEBUG oslo.service.loopingcall [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.062726] env[61991]: DEBUG nova.compute.manager [-] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.062855] env[61991]: DEBUG nova.network.neutron [-] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 860.102310] env[61991]: INFO nova.compute.manager [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Took 31.43 seconds to build instance. [ 860.117309] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.117678] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.117908] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.118760] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.118868] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.119458] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.119458] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.119581] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.119676] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.119919] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.120229] env[61991]: DEBUG nova.virt.hardware [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.121152] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2b5240-f621-4223-b00b-5c2976d34aba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.132635] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db92e5e0-8462-4ffe-9f34-a803ed78a991 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.148200] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:50:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58dc3a16-cc46-440b-b310-026854eda3a2', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.156394] env[61991]: DEBUG oslo.service.loopingcall [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.156725] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 860.156962] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d2e59ec-6ee2-4911-a1c8-cbf54d74b3ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.176777] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.176777] env[61991]: value = "task-1129893" [ 860.176777] env[61991]: _type = "Task" [ 860.176777] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.184438] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129893, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.264351] env[61991]: DEBUG oslo_vmware.api [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129891, 'name': PowerOffVM_Task, 'duration_secs': 0.227302} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.264351] env[61991]: DEBUG nova.network.neutron [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Successfully updated port: f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.265322] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.265519] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.268480] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83c32c50-4d60-4233-bcd2-ada25542658e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.271689] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129892, 'name': CreateVM_Task, 'duration_secs': 0.370949} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.271689] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.272387] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.302087] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523a36e9-b0f0-5f22-87d1-275d396e7ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.013244} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.302298] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.302544] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.302804] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.302985] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.303208] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.303527] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.303851] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.304141] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f74bec7-97d6-47e7-9b83-751af32fc332 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.306107] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe6b4fcd-4a90-487c-a489-2b863640a081 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.311183] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 860.311183] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d32ba5-29f5-3898-4414-35ece38522e4" [ 860.311183] env[61991]: _type = "Task" [ 860.311183] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.316284] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.316487] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.317740] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0660790-0852-4c41-97ae-e00cc4faa32a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.323715] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d32ba5-29f5-3898-4414-35ece38522e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.327228] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 860.327228] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5246cf33-9479-4066-732c-b3e8d266f042" [ 860.327228] env[61991]: _type = "Task" [ 860.327228] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.338785] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5246cf33-9479-4066-732c-b3e8d266f042, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.340064] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.340276] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.340459] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleting the datastore file [datastore2] 7a468ab7-6ca1-43e2-a9b2-89e7640d8148 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.340711] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-171a97b0-d25e-42c7-8807-7ca96c792103 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.346631] env[61991]: DEBUG oslo_vmware.api [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for the task: (returnval){ [ 860.346631] env[61991]: value = "task-1129895" [ 860.346631] env[61991]: _type = "Task" [ 860.346631] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.357042] env[61991]: DEBUG oslo_vmware.api [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.438696] env[61991]: DEBUG nova.compute.manager [req-8bbfa769-4a9a-4e8c-ae01-b07e84152ac6 req-af046fb0-a14f-4edf-b808-e065099c86ad service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Received event network-vif-deleted-87e2f32e-53df-49a9-bae3-f9529f2d3c6c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.438913] env[61991]: INFO nova.compute.manager [req-8bbfa769-4a9a-4e8c-ae01-b07e84152ac6 req-af046fb0-a14f-4edf-b808-e065099c86ad service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Neutron deleted interface 87e2f32e-53df-49a9-bae3-f9529f2d3c6c; detaching it from the instance and deleting it from the info cache [ 860.439223] env[61991]: DEBUG nova.network.neutron [req-8bbfa769-4a9a-4e8c-ae01-b07e84152ac6 req-af046fb0-a14f-4edf-b808-e065099c86ad service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.542265] env[61991]: DEBUG nova.compute.utils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.543757] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.543940] env[61991]: DEBUG nova.network.neutron [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 860.606329] env[61991]: DEBUG oslo_concurrency.lockutils [None req-03ffaedf-7148-43f2-aa55-e566bd98bf1e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.940s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.618566] env[61991]: DEBUG nova.policy [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3005c87b7d245cd8a35fdab04e808e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '537c0b730d2d47498fcf8a8ddd238c3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.688455] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129893, 'name': CreateVM_Task, 'duration_secs': 0.379215} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.690822] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.691732] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.691937] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.692313] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.692888] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aff165c9-3878-4b62-833b-62591d6e786c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.697841] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 860.697841] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52401a75-0004-3161-1efe-08c52c9b5a55" [ 860.697841] env[61991]: _type = "Task" [ 860.697841] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.709573] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52401a75-0004-3161-1efe-08c52c9b5a55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.743936] env[61991]: DEBUG nova.network.neutron [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Updated VIF entry in instance network info cache for port 490686f3-4e55-4f9f-afe0-6dbb260e6bea. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.744872] env[61991]: DEBUG nova.network.neutron [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Updating instance_info_cache with network_info: [{"id": "490686f3-4e55-4f9f-afe0-6dbb260e6bea", "address": "fa:16:3e:cb:98:e2", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap490686f3-4e", "ovs_interfaceid": "490686f3-4e55-4f9f-afe0-6dbb260e6bea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.766875] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.767330] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.767576] env[61991]: DEBUG nova.network.neutron [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.824298] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d32ba5-29f5-3898-4414-35ece38522e4, 'name': SearchDatastore_Task, 'duration_secs': 0.012527} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.824594] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.824828] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.825090] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.839387] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5246cf33-9479-4066-732c-b3e8d266f042, 'name': SearchDatastore_Task, 'duration_secs': 0.014426} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.842823] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0db89863-1b66-4d01-a853-9c8935102ea2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.847750] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 860.847750] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526d51bd-cc25-75b2-561f-e6509ef66a8d" [ 860.847750] env[61991]: _type = "Task" [ 860.847750] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.865272] env[61991]: DEBUG oslo_vmware.api [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Task: {'id': task-1129895, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305276} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.868914] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.868914] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.868914] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.869084] env[61991]: INFO nova.compute.manager [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Took 1.15 seconds to destroy the instance on the hypervisor. [ 860.869243] env[61991]: DEBUG oslo.service.loopingcall [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.869440] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526d51bd-cc25-75b2-561f-e6509ef66a8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.870434] env[61991]: DEBUG nova.compute.manager [-] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.870530] env[61991]: DEBUG nova.network.neutron [-] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 860.872724] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad0167f-e830-4ce6-ab69-a010b85ec454 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.881018] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a3baec-6ea5-416f-802d-123761914a1b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.885112] env[61991]: DEBUG nova.network.neutron [-] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.920777] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62cb3bc5-7c29-4dd0-9fa3-0222ccde8420 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.938480] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc3b75a-8031-4f07-8e36-969740b3bc16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.943296] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dda4217b-66d2-4fa9-99c3-b1820ec1f532 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.954866] env[61991]: DEBUG nova.compute.provider_tree [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.961109] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7423d9c7-7c37-45e1-adce-115453a201c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.980532] env[61991]: DEBUG nova.network.neutron [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Successfully created port: 7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.993520] env[61991]: DEBUG nova.compute.manager [req-8bbfa769-4a9a-4e8c-ae01-b07e84152ac6 req-af046fb0-a14f-4edf-b808-e065099c86ad service nova] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Detach interface failed, port_id=87e2f32e-53df-49a9-bae3-f9529f2d3c6c, reason: Instance c1d4d2eb-810c-42f9-a244-e5aed6e8d517 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 861.050485] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.209255] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52401a75-0004-3161-1efe-08c52c9b5a55, 'name': SearchDatastore_Task, 'duration_secs': 0.010101} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.209255] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.209255] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.210145] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.210145] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.210145] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.210145] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9c13bd8-20f2-4c2a-8db4-122c35ce8d18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.220492] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.220692] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.221413] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b2803b4-dbda-4d59-925a-ddf9c71d32e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.226975] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 861.226975] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba4b00-98c5-bd4e-3d05-91d5e60194b2" [ 861.226975] env[61991]: _type = "Task" [ 861.226975] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.235859] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba4b00-98c5-bd4e-3d05-91d5e60194b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.248964] env[61991]: DEBUG oslo_concurrency.lockutils [req-2518c058-7c9f-4075-9890-f7d6738356db req-bc60eb30-125f-43ee-aa83-b944e2223594 service nova] Releasing lock "refresh_cache-7634e00f-48ee-40a0-8625-2f8dc83cd58d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.298083] env[61991]: DEBUG nova.network.neutron [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.361876] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526d51bd-cc25-75b2-561f-e6509ef66a8d, 'name': SearchDatastore_Task, 'duration_secs': 0.026311} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.362158] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.362419] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/f7eab1a9-2a8a-4664-8736-ca599e207c31.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.362693] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.362877] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.363103] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a06a2404-28d1-4c77-bc3a-25a0136b6ec0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.364843] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ed9f78b-f73a-437f-9637-cf5f5532a20a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.372937] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 861.372937] env[61991]: value = "task-1129896" [ 861.372937] env[61991]: _type = "Task" [ 861.372937] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.374026] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.374026] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.377159] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bc22922-d953-4cce-8d8e-a6c789fa14e1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.384080] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.385371] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 861.385371] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5200a4c6-961a-ffaf-0947-57fd2dd143cc" [ 861.385371] env[61991]: _type = "Task" [ 861.385371] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.390134] env[61991]: INFO nova.compute.manager [-] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Took 1.33 seconds to deallocate network for instance. [ 861.397832] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5200a4c6-961a-ffaf-0947-57fd2dd143cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.434190] env[61991]: DEBUG nova.network.neutron [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.459955] env[61991]: DEBUG nova.scheduler.client.report [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.664978] env[61991]: DEBUG nova.network.neutron [-] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.739401] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba4b00-98c5-bd4e-3d05-91d5e60194b2, 'name': SearchDatastore_Task, 'duration_secs': 0.039953} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.740302] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3b413fc-10ee-4723-90c4-1efe8d1f3833 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.748981] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 861.748981] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526353eb-3f08-1957-7850-9f49f5dcbc1d" [ 861.748981] env[61991]: _type = "Task" [ 861.748981] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.759569] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526353eb-3f08-1957-7850-9f49f5dcbc1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.883067] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129896, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.897318] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5200a4c6-961a-ffaf-0947-57fd2dd143cc, 'name': SearchDatastore_Task, 'duration_secs': 0.012282} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.899397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.899654] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9ea8630-4766-4bf8-a378-bbe4cbe9f826 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.905515] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 861.905515] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c57c11-ad27-b76a-031b-49b6b34e5708" [ 861.905515] env[61991]: _type = "Task" [ 861.905515] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.915464] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c57c11-ad27-b76a-031b-49b6b34e5708, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.938031] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.938031] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Instance network_info: |[{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.938978] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:73:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3f695b6-65bc-45cc-a61d-3c38a14e5c0c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f07a2552-10f4-48c4-9a79-b66506f34302', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.946463] env[61991]: DEBUG oslo.service.loopingcall [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.946543] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.946806] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2afc3536-14df-489e-a753-11ae6eeeb173 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.964976] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.968582] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 26.633s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.971603] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.971603] env[61991]: value = "task-1129897" [ 861.971603] env[61991]: _type = "Task" [ 861.971603] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.978886] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129897, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.991272] env[61991]: INFO nova.scheduler.client.report [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocations for instance 15215bba-8160-408b-9e8f-afbb9e71d50e [ 862.060408] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.084592] env[61991]: DEBUG nova.compute.manager [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-vif-plugged-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.084822] env[61991]: DEBUG oslo_concurrency.lockutils [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.085100] env[61991]: DEBUG oslo_concurrency.lockutils [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.085262] env[61991]: DEBUG oslo_concurrency.lockutils [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.085423] env[61991]: DEBUG nova.compute.manager [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] No waiting events found dispatching network-vif-plugged-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.085593] env[61991]: WARNING nova.compute.manager [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received unexpected event network-vif-plugged-f07a2552-10f4-48c4-9a79-b66506f34302 for instance with vm_state building and task_state spawning. [ 862.085754] env[61991]: DEBUG nova.compute.manager [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.085908] env[61991]: DEBUG nova.compute.manager [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing instance network info cache due to event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.086104] env[61991]: DEBUG oslo_concurrency.lockutils [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.086315] env[61991]: DEBUG oslo_concurrency.lockutils [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.086434] env[61991]: DEBUG nova.network.neutron [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 862.089814] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.090044] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.090215] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.090406] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.090553] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.090700] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.090908] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.091086] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.091283] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.091423] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.091602] env[61991]: DEBUG nova.virt.hardware [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.092710] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dcf37a-9594-4849-992e-4de0f259ade3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.104061] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194fa456-0b79-4115-816c-01d9f65c57ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.167802] env[61991]: INFO nova.compute.manager [-] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Took 1.30 seconds to deallocate network for instance. [ 862.259137] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526353eb-3f08-1957-7850-9f49f5dcbc1d, 'name': SearchDatastore_Task, 'duration_secs': 0.017996} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.259435] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.259735] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.260008] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ac281e8-2de7-4472-9612-39d57e518c9c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.267043] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 862.267043] env[61991]: value = "task-1129898" [ 862.267043] env[61991]: _type = "Task" [ 862.267043] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.275276] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.350410] env[61991]: DEBUG nova.network.neutron [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updated VIF entry in instance network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 862.350803] env[61991]: DEBUG nova.network.neutron [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.383489] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581116} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.383761] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/f7eab1a9-2a8a-4664-8736-ca599e207c31.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.383982] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.384289] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d2ee35a-2302-4cce-bbb1-d2a37197cb57 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.390185] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 862.390185] env[61991]: value = "task-1129899" [ 862.390185] env[61991]: _type = "Task" [ 862.390185] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.397811] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.415035] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c57c11-ad27-b76a-031b-49b6b34e5708, 'name': SearchDatastore_Task, 'duration_secs': 0.027357} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.415372] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.415644] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 7634e00f-48ee-40a0-8625-2f8dc83cd58d/7634e00f-48ee-40a0-8625-2f8dc83cd58d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.415901] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50bfeb8c-03c4-4cdd-a129-67fefeba14d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.421888] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 862.421888] env[61991]: value = "task-1129900" [ 862.421888] env[61991]: _type = "Task" [ 862.421888] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.430471] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.468281] env[61991]: DEBUG nova.compute.manager [req-e258ef2d-592f-4169-80c9-1125a3ac7bfb req-ca95d91c-34cb-4df9-bb5d-79f8d6b7343c service nova] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Received event network-vif-deleted-84517dfd-59cc-45de-bde0-638279ae93c8 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.473559] env[61991]: INFO nova.compute.claims [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.489696] env[61991]: DEBUG nova.network.neutron [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Successfully updated port: 7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.490934] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129897, 'name': CreateVM_Task, 'duration_secs': 0.33735} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.491117] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.493373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.493373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.493373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.493373] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ce9dfcc-20c2-4680-94c9-4bcdd5b98276 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.498767] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6f45abda-7550-43af-b9b9-3d750710e9e8 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "15215bba-8160-408b-9e8f-afbb9e71d50e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.334s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.503779] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 862.503779] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527f4b77-5baa-334e-75d4-3bac15945f41" [ 862.503779] env[61991]: _type = "Task" [ 862.503779] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.512845] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527f4b77-5baa-334e-75d4-3bac15945f41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.675364] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.778614] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129898, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.853514] env[61991]: DEBUG oslo_concurrency.lockutils [req-7dbe7a01-d716-4496-a19c-fdaf3e90df16 req-db050af9-0b52-4bd4-ab5b-2876edb47e88 service nova] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.901328] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062308} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.901643] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.902493] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d46be61-4fce-49ed-9ba8-d61c5261fc4a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.926829] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/f7eab1a9-2a8a-4664-8736-ca599e207c31.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.927176] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0306361b-e2a8-4b4a-ab11-8a836aebae94 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.951996] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129900, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.953714] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 862.953714] env[61991]: value = "task-1129901" [ 862.953714] env[61991]: _type = "Task" [ 862.953714] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.964660] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129901, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.985973] env[61991]: INFO nova.compute.resource_tracker [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating resource usage from migration f8d56f1f-b98b-44cc-a80d-b069ddd2017b [ 862.994123] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "refresh_cache-d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.994299] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired lock "refresh_cache-d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.994420] env[61991]: DEBUG nova.network.neutron [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 863.015868] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527f4b77-5baa-334e-75d4-3bac15945f41, 'name': SearchDatastore_Task, 'duration_secs': 0.023732} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.019228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.019536] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.019779] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.019938] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.020143] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.021033] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-213af142-59a9-404b-985c-ac4cc488a88c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.039029] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.039141] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.039961] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-547db0e4-b35f-4682-9a46-176696a055fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.050125] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 863.050125] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dd8193-31d9-cae9-3fb9-43ec682a6bdb" [ 863.050125] env[61991]: _type = "Task" [ 863.050125] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.064285] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dd8193-31d9-cae9-3fb9-43ec682a6bdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.285772] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129898, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.826327} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.288101] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.288748] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.289039] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27adf2be-9e83-4596-a3df-7da748715d4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.297113] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 863.297113] env[61991]: value = "task-1129902" [ 863.297113] env[61991]: _type = "Task" [ 863.297113] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.309976] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129902, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.316466] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218d522c-977f-4857-9918-7d8c7b9ace16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.325130] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded24da0-e9b3-49c0-b7f2-662f224cb5ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.357880] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e8674f-a962-40c2-aa20-b98e607fb81e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.365848] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6b827c-9ce7-4cb1-95de-4d82c9c9ca3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.379661] env[61991]: DEBUG nova.compute.provider_tree [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.438371] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.949515} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.438647] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 7634e00f-48ee-40a0-8625-2f8dc83cd58d/7634e00f-48ee-40a0-8625-2f8dc83cd58d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.438863] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.439140] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91222efe-f0ab-4366-8bb4-22a1181d0aab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.445657] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 863.445657] env[61991]: value = "task-1129903" [ 863.445657] env[61991]: _type = "Task" [ 863.445657] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.453972] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129903, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.463967] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129901, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.527793] env[61991]: DEBUG nova.network.neutron [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.559604] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dd8193-31d9-cae9-3fb9-43ec682a6bdb, 'name': SearchDatastore_Task, 'duration_secs': 0.034074} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.560429] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e820205d-1dd6-4b90-af4c-36a32f9b9fd4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.565308] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 863.565308] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f65946-6130-50fb-3f93-5d1ee8d1d578" [ 863.565308] env[61991]: _type = "Task" [ 863.565308] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.574564] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f65946-6130-50fb-3f93-5d1ee8d1d578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.663713] env[61991]: DEBUG nova.network.neutron [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Updating instance_info_cache with network_info: [{"id": "7049a455-7d87-4402-a422-05ce15c9264b", "address": "fa:16:3e:44:bc:d6", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7049a455-7d", "ovs_interfaceid": "7049a455-7d87-4402-a422-05ce15c9264b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.798322] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "39004ce4-c96b-4911-9b64-3c508ddcecaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.798675] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.810602] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129902, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.190533} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.812151] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.812151] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb36a14e-f7e6-4cfc-a702-1c99a1917b6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.834603] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.835070] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da713507-9d6e-457a-9893-473a06b215c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.855682] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 863.855682] env[61991]: value = "task-1129904" [ 863.855682] env[61991]: _type = "Task" [ 863.855682] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.862863] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129904, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.883099] env[61991]: DEBUG nova.scheduler.client.report [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.955617] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129903, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099272} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.958539] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.959274] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7610e3-9be0-4074-a1c7-d73f6e4076cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.965900] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129901, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.982930] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 7634e00f-48ee-40a0-8625-2f8dc83cd58d/7634e00f-48ee-40a0-8625-2f8dc83cd58d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.983138] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e402fc6-a36d-4cd4-a0b3-b4813045dab7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.002931] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 864.002931] env[61991]: value = "task-1129905" [ 864.002931] env[61991]: _type = "Task" [ 864.002931] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.010974] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129905, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.074927] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f65946-6130-50fb-3f93-5d1ee8d1d578, 'name': SearchDatastore_Task, 'duration_secs': 0.084346} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.075259] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.075536] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 9d1ac179-3d7e-4f0b-a80d-033eba447cab/9d1ac179-3d7e-4f0b-a80d-033eba447cab.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.076139] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88bdf3b5-a1e8-4ea8-a57e-697e8ae9b571 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.081761] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 864.081761] env[61991]: value = "task-1129906" [ 864.081761] env[61991]: _type = "Task" [ 864.081761] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.089151] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.114787] env[61991]: DEBUG nova.compute.manager [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Received event network-vif-plugged-7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.115106] env[61991]: DEBUG oslo_concurrency.lockutils [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] Acquiring lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.115416] env[61991]: DEBUG oslo_concurrency.lockutils [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.115624] env[61991]: DEBUG oslo_concurrency.lockutils [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.115818] env[61991]: DEBUG nova.compute.manager [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] No waiting events found dispatching network-vif-plugged-7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.115992] env[61991]: WARNING nova.compute.manager [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Received unexpected event network-vif-plugged-7049a455-7d87-4402-a422-05ce15c9264b for instance with vm_state building and task_state spawning. [ 864.116194] env[61991]: DEBUG nova.compute.manager [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Received event network-changed-7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.116351] env[61991]: DEBUG nova.compute.manager [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Refreshing instance network info cache due to event network-changed-7049a455-7d87-4402-a422-05ce15c9264b. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.116547] env[61991]: DEBUG oslo_concurrency.lockutils [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] Acquiring lock "refresh_cache-d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.166591] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Releasing lock "refresh_cache-d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.166962] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Instance network_info: |[{"id": "7049a455-7d87-4402-a422-05ce15c9264b", "address": "fa:16:3e:44:bc:d6", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7049a455-7d", "ovs_interfaceid": "7049a455-7d87-4402-a422-05ce15c9264b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 864.167305] env[61991]: DEBUG oslo_concurrency.lockutils [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] Acquired lock "refresh_cache-d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.167533] env[61991]: DEBUG nova.network.neutron [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Refreshing network info cache for port 7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.170040] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:bc:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9732690c-bdcf-4e6f-9a32-42c196333eb8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7049a455-7d87-4402-a422-05ce15c9264b', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.176525] env[61991]: DEBUG oslo.service.loopingcall [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.177430] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.177677] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82822eb8-91f4-4469-b820-af0e1dd072a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.199588] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.199588] env[61991]: value = "task-1129907" [ 864.199588] env[61991]: _type = "Task" [ 864.199588] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.206549] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129907, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.306057] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.366646] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129904, 'name': ReconfigVM_Task, 'duration_secs': 0.306629} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.367072] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 3183190e-3202-40be-894c-e47b186708b5/3183190e-3202-40be-894c-e47b186708b5.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.367872] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93f3baf0-4a2d-487a-9bab-856fa07db142 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.375116] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 864.375116] env[61991]: value = "task-1129908" [ 864.375116] env[61991]: _type = "Task" [ 864.375116] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.383951] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129908, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.387803] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.419s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.387998] env[61991]: INFO nova.compute.manager [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Migrating [ 864.395059] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.297s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.396582] env[61991]: INFO nova.compute.claims [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.467211] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129901, 'name': ReconfigVM_Task, 'duration_secs': 1.199984} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.467547] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Reconfigured VM instance instance-0000004e to attach disk [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/f7eab1a9-2a8a-4664-8736-ca599e207c31.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.468258] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5109a3e0-f896-4c01-a421-ac2bf55eb1ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.479019] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 864.479019] env[61991]: value = "task-1129909" [ 864.479019] env[61991]: _type = "Task" [ 864.479019] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.490111] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129909, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.514280] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129905, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.593317] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129906, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.712060] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129907, 'name': CreateVM_Task, 'duration_secs': 0.369943} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.714887] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.715696] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.715868] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.716210] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.716793] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8129231-af3b-41de-9f27-902dccae1a7b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.723456] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 864.723456] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52587434-6cd2-e787-25d4-ad838708e0e8" [ 864.723456] env[61991]: _type = "Task" [ 864.723456] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.731825] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52587434-6cd2-e787-25d4-ad838708e0e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.829676] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.883955] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129908, 'name': Rename_Task, 'duration_secs': 0.165063} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.884294] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.884557] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7062398-0600-46d0-a1be-9aa43b05f060 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.891626] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 864.891626] env[61991]: value = "task-1129910" [ 864.891626] env[61991]: _type = "Task" [ 864.891626] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.899776] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129910, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.909075] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.909278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.909457] env[61991]: DEBUG nova.network.neutron [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.913056] env[61991]: DEBUG nova.network.neutron [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Updated VIF entry in instance network info cache for port 7049a455-7d87-4402-a422-05ce15c9264b. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.913415] env[61991]: DEBUG nova.network.neutron [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Updating instance_info_cache with network_info: [{"id": "7049a455-7d87-4402-a422-05ce15c9264b", "address": "fa:16:3e:44:bc:d6", "network": {"id": "8bcad8da-3d64-46a6-aa8b-027d8bd90e10", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.222", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "351f52655e4f4262b51dd57762a56c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9732690c-bdcf-4e6f-9a32-42c196333eb8", "external-id": "nsx-vlan-transportzone-548", "segmentation_id": 548, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7049a455-7d", "ovs_interfaceid": "7049a455-7d87-4402-a422-05ce15c9264b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.988795] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129909, 'name': Rename_Task, 'duration_secs': 0.241439} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.989193] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.989490] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b978ddeb-5d08-4bd4-83e9-318fbd9920d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.995484] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 864.995484] env[61991]: value = "task-1129911" [ 864.995484] env[61991]: _type = "Task" [ 864.995484] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.003291] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.011653] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129905, 'name': ReconfigVM_Task, 'duration_secs': 0.547852} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.011811] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 7634e00f-48ee-40a0-8625-2f8dc83cd58d/7634e00f-48ee-40a0-8625-2f8dc83cd58d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.012441] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84a15a27-488c-4edf-8587-ccb6a9afab22 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.017915] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 865.017915] env[61991]: value = "task-1129912" [ 865.017915] env[61991]: _type = "Task" [ 865.017915] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.025563] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129912, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.091833] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129906, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.233402] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52587434-6cd2-e787-25d4-ad838708e0e8, 'name': SearchDatastore_Task, 'duration_secs': 0.028514} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.233721] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.233959] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.234212] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.234363] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.234544] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.234799] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48b75684-3bb9-437d-9494-636d9e31701a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.245912] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.246103] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.246819] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f896da87-49b0-4f6e-a728-aef3b86103d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.251577] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 865.251577] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528fab57-0b03-0c80-75af-cc44547131e5" [ 865.251577] env[61991]: _type = "Task" [ 865.251577] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.258704] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528fab57-0b03-0c80-75af-cc44547131e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.399731] env[61991]: DEBUG oslo_vmware.api [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129910, 'name': PowerOnVM_Task, 'duration_secs': 0.504355} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.400020] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.400244] env[61991]: DEBUG nova.compute.manager [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.401101] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fee484-edcf-496c-8048-4133bc18765e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.416509] env[61991]: DEBUG oslo_concurrency.lockutils [req-bc022e56-e0ee-4277-bc6c-895ee6dce3c0 req-c21b4a34-2292-4a75-bc11-0d9df3eb3789 service nova] Releasing lock "refresh_cache-d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.508717] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129911, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.527591] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129912, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.593032] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129906, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.708014] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2285021d-1f71-4f75-8b8f-ae65a330ebca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.716530] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decbf91c-eb7c-4cc7-b37c-994afd374223 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.722182] env[61991]: DEBUG nova.network.neutron [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.758408] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4b3516-f980-4768-b181-7e577a9ab2e1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.769190] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0fc9fa-155d-4f80-97b6-a679cacb4a81 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.772854] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528fab57-0b03-0c80-75af-cc44547131e5, 'name': SearchDatastore_Task, 'duration_secs': 0.008258} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.774050] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe0ea0ff-3f7a-43c0-ba43-c16258d0255a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.783483] env[61991]: DEBUG nova.compute.provider_tree [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.787741] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 865.787741] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d58cae-1c57-ce1d-aec2-ae733ea79f66" [ 865.787741] env[61991]: _type = "Task" [ 865.787741] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.795318] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d58cae-1c57-ce1d-aec2-ae733ea79f66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.917088] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.006949] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129911, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.027396] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129912, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.094284] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129906, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.704028} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.094557] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 9d1ac179-3d7e-4f0b-a80d-033eba447cab/9d1ac179-3d7e-4f0b-a80d-033eba447cab.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.094797] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.095064] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f286dcfe-245c-4674-8d66-363d1c043bc8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.101710] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 866.101710] env[61991]: value = "task-1129913" [ 866.101710] env[61991]: _type = "Task" [ 866.101710] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.109535] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.227044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.286574] env[61991]: DEBUG nova.scheduler.client.report [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.301977] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d58cae-1c57-ce1d-aec2-ae733ea79f66, 'name': SearchDatastore_Task, 'duration_secs': 0.013245} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.302278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.302675] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3/d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.303206] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55301c06-dda0-4442-8ee2-a2058520db7b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.310514] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 866.310514] env[61991]: value = "task-1129914" [ 866.310514] env[61991]: _type = "Task" [ 866.310514] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.318088] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.508726] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129911, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.529542] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129912, 'name': Rename_Task, 'duration_secs': 1.237271} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.529879] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.530150] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e95cd930-1386-443f-9baf-18c9a9ee66a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.537222] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 866.537222] env[61991]: value = "task-1129915" [ 866.537222] env[61991]: _type = "Task" [ 866.537222] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.548783] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.613086] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066971} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.613474] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.614617] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2683b1-7346-4a3d-a27a-2794399d9006 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.640449] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 9d1ac179-3d7e-4f0b-a80d-033eba447cab/9d1ac179-3d7e-4f0b-a80d-033eba447cab.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.640870] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb59949b-0445-41a4-b5de-caf47d44d2e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.661729] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 866.661729] env[61991]: value = "task-1129916" [ 866.661729] env[61991]: _type = "Task" [ 866.661729] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.672215] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129916, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.775143] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "3183190e-3202-40be-894c-e47b186708b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.776220] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "3183190e-3202-40be-894c-e47b186708b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.776730] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "3183190e-3202-40be-894c-e47b186708b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.777371] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "3183190e-3202-40be-894c-e47b186708b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.779052] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "3183190e-3202-40be-894c-e47b186708b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.780486] env[61991]: INFO nova.compute.manager [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Terminating instance [ 866.782643] env[61991]: DEBUG nova.compute.manager [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.783166] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.784211] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d1dc13-6bac-4d09-a627-bceefbbb8593 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.793243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.793907] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.797402] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.798087] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.451s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.798466] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.798863] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 866.799329] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.929s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.799757] env[61991]: DEBUG nova.objects.instance [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'resources' on Instance uuid 62b3e5ae-b90c-47f8-95c4-14587dbf647d {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.801627] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8e7901c-77d5-4e3e-b15c-b5698d5ca45b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.805077] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42daf5a1-39e8-430c-9d94-763c60bc9c77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.821196] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79901ea8-7b6e-4a10-bab2-0bc9e4daee68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.826736] env[61991]: DEBUG oslo_vmware.api [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 866.826736] env[61991]: value = "task-1129917" [ 866.826736] env[61991]: _type = "Task" [ 866.826736] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.845969] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129914, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491236} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.847040] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3/d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.847325] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.848109] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85204c2c-7186-4e28-b0f8-671c8732cd9f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.853850] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b337042-1f55-4b30-a017-affd80989df8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.856360] env[61991]: DEBUG oslo_vmware.api [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.863107] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed302771-d609-464a-83ef-0f5adbeff9be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.866422] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 866.866422] env[61991]: value = "task-1129918" [ 866.866422] env[61991]: _type = "Task" [ 866.866422] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.898867] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180263MB free_disk=167GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 866.898867] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.905018] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.007570] env[61991]: DEBUG oslo_vmware.api [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129911, 'name': PowerOnVM_Task, 'duration_secs': 1.612439} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.007875] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.008680] env[61991]: INFO nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Took 12.38 seconds to spawn the instance on the hypervisor. [ 867.008680] env[61991]: DEBUG nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.009196] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b284c1c-073c-4646-9758-906b03625a4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.047202] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129915, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.172762] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129916, 'name': ReconfigVM_Task, 'duration_secs': 0.402946} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.173084] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 9d1ac179-3d7e-4f0b-a80d-033eba447cab/9d1ac179-3d7e-4f0b-a80d-033eba447cab.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.173770] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7649d6e-84ea-4f5a-b482-e67c5642af49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.180054] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 867.180054] env[61991]: value = "task-1129919" [ 867.180054] env[61991]: _type = "Task" [ 867.180054] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.188056] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129919, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.299463] env[61991]: DEBUG nova.compute.utils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.300919] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.301118] env[61991]: DEBUG nova.network.neutron [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 867.341031] env[61991]: DEBUG oslo_vmware.api [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129917, 'name': PowerOffVM_Task, 'duration_secs': 0.212012} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.343497] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.343676] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.345150] env[61991]: DEBUG nova.policy [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc2e1f75d6bd44df96ad5b229d883b31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25647cd0715f490092623f4ead833738', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.346465] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e206ece1-7161-46aa-aa63-6f44a499407a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.376141] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072411} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.376492] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.377285] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c050ce4e-e06f-4278-9907-1796192f4cd5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.399504] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3/d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.402845] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49018c83-e284-44d1-9332-dfc421a58d30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.418920] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.419252] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.419369] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore1] 3183190e-3202-40be-894c-e47b186708b5 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.419969] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89f9be11-bf00-4d5a-b831-5399f0fbe12c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.424276] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 867.424276] env[61991]: value = "task-1129921" [ 867.424276] env[61991]: _type = "Task" [ 867.424276] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.428672] env[61991]: DEBUG oslo_vmware.api [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 867.428672] env[61991]: value = "task-1129922" [ 867.428672] env[61991]: _type = "Task" [ 867.428672] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.437719] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129921, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.442674] env[61991]: DEBUG oslo_vmware.api [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129922, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.494336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "0dcf415e-b554-4c82-b127-9f0e22176031" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.494336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "0dcf415e-b554-4c82-b127-9f0e22176031" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.530186] env[61991]: INFO nova.compute.manager [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Took 38.36 seconds to build instance. [ 867.553511] env[61991]: DEBUG oslo_vmware.api [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129915, 'name': PowerOnVM_Task, 'duration_secs': 0.567684} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.553752] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.553971] env[61991]: INFO nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Took 10.44 seconds to spawn the instance on the hypervisor. [ 867.554181] env[61991]: DEBUG nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.555258] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb08117-f31c-4fcb-9b8e-47438e3fa289 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.645418] env[61991]: DEBUG nova.network.neutron [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Successfully created port: 3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.674331] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c4beda-c32e-4917-a67f-45dae3b22331 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.685015] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc648b0f-8e3c-4b7c-be21-eac9a47181ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.694767] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129919, 'name': Rename_Task, 'duration_secs': 0.145712} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.723240] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.723240] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ceae28f-b06c-4012-93d3-97467dc6c0b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.724812] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd133ef7-54f7-4b72-b6b6-8f7de7b23e29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.733102] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97522840-76f8-4918-908d-19637c2858e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.737165] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 867.737165] env[61991]: value = "task-1129923" [ 867.737165] env[61991]: _type = "Task" [ 867.737165] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.752577] env[61991]: DEBUG nova.compute.provider_tree [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 867.754795] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0623c8-8fe5-40e0-9d25-05eff55031d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.762618] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129923, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.775920] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 867.804387] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.937881] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129921, 'name': ReconfigVM_Task, 'duration_secs': 0.291754} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.938618] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Reconfigured VM instance instance-00000051 to attach disk [datastore1] d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3/d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.939323] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a51da911-e58c-4d88-a524-cb072c2a9b2b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.943648] env[61991]: DEBUG oslo_vmware.api [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129922, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316416} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.944968] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.944968] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.944968] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.944968] env[61991]: INFO nova.compute.manager [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 867.945139] env[61991]: DEBUG oslo.service.loopingcall [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.945317] env[61991]: DEBUG nova.compute.manager [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.945416] env[61991]: DEBUG nova.network.neutron [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 867.951789] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 867.951789] env[61991]: value = "task-1129924" [ 867.951789] env[61991]: _type = "Task" [ 867.951789] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.960040] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129924, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.998179] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 868.035478] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a06044-4276-497f-875b-a887ce7a1d36 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.875s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.076245] env[61991]: INFO nova.compute.manager [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Took 36.94 seconds to build instance. [ 868.251216] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129923, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.274342] env[61991]: ERROR nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [req-be346cde-59dd-4c06-859d-81a1d406c0e8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-be346cde-59dd-4c06-859d-81a1d406c0e8"}]} [ 868.281135] env[61991]: DEBUG nova.compute.manager [req-58fc7361-d21d-4940-a0ca-950631dd22c7 req-11ea2c3a-3892-483e-8121-9fa79b41dd8e service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Received event network-vif-deleted-58dc3a16-cc46-440b-b310-026854eda3a2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.281135] env[61991]: INFO nova.compute.manager [req-58fc7361-d21d-4940-a0ca-950631dd22c7 req-11ea2c3a-3892-483e-8121-9fa79b41dd8e service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Neutron deleted interface 58dc3a16-cc46-440b-b310-026854eda3a2; detaching it from the instance and deleting it from the info cache [ 868.281135] env[61991]: DEBUG nova.network.neutron [req-58fc7361-d21d-4940-a0ca-950631dd22c7 req-11ea2c3a-3892-483e-8121-9fa79b41dd8e service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.282297] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.282718] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc4b1bba-5ef8-4ff8-a06c-f092458ea9a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.289737] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 868.289737] env[61991]: value = "task-1129925" [ 868.289737] env[61991]: _type = "Task" [ 868.289737] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.290653] env[61991]: DEBUG nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 868.301969] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 868.302078] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 868.306051] env[61991]: DEBUG nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 868.306279] env[61991]: DEBUG nova.compute.provider_tree [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 868.323602] env[61991]: DEBUG nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 868.347020] env[61991]: DEBUG nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 868.462803] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129924, 'name': Rename_Task, 'duration_secs': 0.1635} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.465150] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.465690] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a32a6014-0671-4d3e-80f8-5ea169c5a1d6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.471410] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 868.471410] env[61991]: value = "task-1129926" [ 868.471410] env[61991]: _type = "Task" [ 868.471410] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.481248] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129926, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.522317] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.578373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d0fc61e0-238b-4792-ba68-6cc6462f38c9 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.854s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.619522] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e9985b-2700-4f2c-a2f5-19113de0d6b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.626953] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a83a2f-02b2-47ea-a474-0793ba33a74d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.657977] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93110232-a504-4295-8d4d-87038d685ed8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.666497] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24727dad-82fd-47df-ad0a-66ded8204274 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.680441] env[61991]: DEBUG nova.compute.provider_tree [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 868.719238] env[61991]: DEBUG nova.network.neutron [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.748374] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129923, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.784350] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7795be7-9dc9-4287-a0f1-f224128f4295 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.793714] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60632fa-bbb3-4c30-81f8-393f10c9ad95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.814048] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.814312] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.814474] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.814667] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.814816] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.814965] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.815417] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.815594] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.815766] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.815931] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.816130] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.832640] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.834857] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-731522ec-c778-48ee-a16f-b275ff6846f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.846310] env[61991]: DEBUG nova.compute.manager [req-58fc7361-d21d-4940-a0ca-950631dd22c7 req-11ea2c3a-3892-483e-8121-9fa79b41dd8e service nova] [instance: 3183190e-3202-40be-894c-e47b186708b5] Detach interface failed, port_id=58dc3a16-cc46-440b-b310-026854eda3a2, reason: Instance 3183190e-3202-40be-894c-e47b186708b5 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 868.852203] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 868.852203] env[61991]: value = "task-1129927" [ 868.852203] env[61991]: _type = "Task" [ 868.852203] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.861715] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129927, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.871796] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.872073] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.872242] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.872432] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.872676] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.872769] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.873069] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.873310] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.873517] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.873774] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.873973] env[61991]: DEBUG nova.virt.hardware [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.874928] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31be99a7-0617-47a3-8ee3-516571c92b06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.883104] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2a4903-9a70-4d36-8ae9-a1ffde998b5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.905705] env[61991]: INFO nova.compute.manager [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Rescuing [ 868.905911] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.906084] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.906256] env[61991]: DEBUG nova.network.neutron [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.982898] env[61991]: DEBUG oslo_vmware.api [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129926, 'name': PowerOnVM_Task, 'duration_secs': 0.465097} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.983202] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.983438] env[61991]: INFO nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Took 6.92 seconds to spawn the instance on the hypervisor. [ 868.983602] env[61991]: DEBUG nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.984410] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c3445d-9cbb-41f7-a4af-2804bc674085 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.221429] env[61991]: DEBUG nova.compute.manager [req-348348b2-aacb-44ae-9264-05dbb516c0fd req-727601b3-eaba-45b5-9f1f-415f8818b441 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Received event network-vif-plugged-3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.221429] env[61991]: DEBUG oslo_concurrency.lockutils [req-348348b2-aacb-44ae-9264-05dbb516c0fd req-727601b3-eaba-45b5-9f1f-415f8818b441 service nova] Acquiring lock "97913aea-48c1-4fda-aee3-578e89a08f71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.221429] env[61991]: DEBUG oslo_concurrency.lockutils [req-348348b2-aacb-44ae-9264-05dbb516c0fd req-727601b3-eaba-45b5-9f1f-415f8818b441 service nova] Lock "97913aea-48c1-4fda-aee3-578e89a08f71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.221429] env[61991]: DEBUG oslo_concurrency.lockutils [req-348348b2-aacb-44ae-9264-05dbb516c0fd req-727601b3-eaba-45b5-9f1f-415f8818b441 service nova] Lock "97913aea-48c1-4fda-aee3-578e89a08f71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.221429] env[61991]: DEBUG nova.compute.manager [req-348348b2-aacb-44ae-9264-05dbb516c0fd req-727601b3-eaba-45b5-9f1f-415f8818b441 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] No waiting events found dispatching network-vif-plugged-3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.221429] env[61991]: WARNING nova.compute.manager [req-348348b2-aacb-44ae-9264-05dbb516c0fd req-727601b3-eaba-45b5-9f1f-415f8818b441 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Received unexpected event network-vif-plugged-3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 for instance with vm_state building and task_state spawning. [ 869.222434] env[61991]: DEBUG nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 869.222673] env[61991]: DEBUG nova.compute.provider_tree [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 107 to 108 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 869.222857] env[61991]: DEBUG nova.compute.provider_tree [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.228183] env[61991]: INFO nova.compute.manager [-] [instance: 3183190e-3202-40be-894c-e47b186708b5] Took 1.28 seconds to deallocate network for instance. [ 869.250586] env[61991]: DEBUG oslo_vmware.api [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1129923, 'name': PowerOnVM_Task, 'duration_secs': 1.033406} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.250692] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.250871] env[61991]: INFO nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Took 9.63 seconds to spawn the instance on the hypervisor. [ 869.251115] env[61991]: DEBUG nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.251997] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a3ccdc-0e2d-41c9-ae3d-e5e9e96afe1a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.300351] env[61991]: DEBUG nova.network.neutron [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Successfully updated port: 3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.362614] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129927, 'name': ReconfigVM_Task, 'duration_secs': 0.276525} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.362614] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 869.501708] env[61991]: INFO nova.compute.manager [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Took 37.85 seconds to build instance. [ 869.656654] env[61991]: DEBUG nova.network.neutron [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Updating instance_info_cache with network_info: [{"id": "ce8085ee-47de-4c70-aafc-dcdf66c17383", "address": "fa:16:3e:3c:85:e5", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce8085ee-47", "ovs_interfaceid": "ce8085ee-47de-4c70-aafc-dcdf66c17383", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.729359] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.930s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.732599] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.696s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.734038] env[61991]: INFO nova.compute.claims [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.738167] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.752101] env[61991]: INFO nova.scheduler.client.report [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted allocations for instance 62b3e5ae-b90c-47f8-95c4-14587dbf647d [ 869.769360] env[61991]: INFO nova.compute.manager [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Took 38.12 seconds to build instance. [ 869.803100] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "refresh_cache-97913aea-48c1-4fda-aee3-578e89a08f71" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.803293] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "refresh_cache-97913aea-48c1-4fda-aee3-578e89a08f71" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.803538] env[61991]: DEBUG nova.network.neutron [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.870193] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.870505] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.870688] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.870923] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.871116] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.871315] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.871566] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.871777] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.871985] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.872212] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.872424] env[61991]: DEBUG nova.virt.hardware [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.878258] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Reconfiguring VM instance instance-00000033 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 869.879408] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28bff21a-b0c4-4120-a05d-5d34531f8967 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.898241] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 869.898241] env[61991]: value = "task-1129928" [ 869.898241] env[61991]: _type = "Task" [ 869.898241] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.907053] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129928, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.003872] env[61991]: DEBUG oslo_concurrency.lockutils [None req-67177eef-6475-4afb-a3eb-f6b6a4e7513b tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.364s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.158556] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-f7eab1a9-2a8a-4664-8736-ca599e207c31" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.266475] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dabaeeea-cd46-4a36-97a7-154726c98834 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "62b3e5ae-b90c-47f8-95c4-14587dbf647d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.985s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.271432] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4ce43c95-9671-449d-bd04-f701345baf2e tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.642s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.339936] env[61991]: DEBUG nova.network.neutron [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.409021] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129928, 'name': ReconfigVM_Task, 'duration_secs': 0.233076} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.409255] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Reconfigured VM instance instance-00000033 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 870.409991] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e3b2ff-b5e9-4689-b2ab-9c5cd718eb12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.433012] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556/3bcde671-5702-4b8a-8881-88eb7dfd0556.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.433374] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2e18159-a00d-4b43-91a1-05a69ba2d67d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.450977] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 870.450977] env[61991]: value = "task-1129929" [ 870.450977] env[61991]: _type = "Task" [ 870.450977] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.459146] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129929, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.694606] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.695087] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-beb81cfd-3d01-4385-a5c6-785f7f88ce01 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.704603] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 870.704603] env[61991]: value = "task-1129930" [ 870.704603] env[61991]: _type = "Task" [ 870.704603] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.714360] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129930, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.821418] env[61991]: DEBUG nova.network.neutron [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Updating instance_info_cache with network_info: [{"id": "3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63", "address": "fa:16:3e:43:c6:46", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c71ee4b-db", "ovs_interfaceid": "3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.963658] env[61991]: DEBUG oslo_vmware.api [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1129929, 'name': ReconfigVM_Task, 'duration_secs': 0.393124} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.964273] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556/3bcde671-5702-4b8a-8881-88eb7dfd0556.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.964649] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 871.064577] env[61991]: DEBUG nova.compute.manager [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.064818] env[61991]: DEBUG nova.compute.manager [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing instance network info cache due to event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.065120] env[61991]: DEBUG oslo_concurrency.lockutils [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.065549] env[61991]: DEBUG oslo_concurrency.lockutils [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.065708] env[61991]: DEBUG nova.network.neutron [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.068579] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f01d3aa-ec6c-4877-8db0-5fcd548a54c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.077124] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa267c92-afb7-44a5-853a-174d72197feb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.110689] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bbd06a-4d54-415e-8c30-0e3d1a57a504 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.118167] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6256a95-bd6f-4c5d-a398-180f4ab5e984 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.133503] env[61991]: DEBUG nova.compute.provider_tree [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.216964] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129930, 'name': PowerOffVM_Task, 'duration_secs': 0.220598} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.217278] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 871.218417] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca72bb97-0058-4709-a453-d020bccae1dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.238120] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51eca14-1227-4d0c-8554-3e936c56bfad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.270056] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.270436] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9897f20c-55f1-402c-b5d3-7831410f06a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.277709] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 871.277709] env[61991]: value = "task-1129931" [ 871.277709] env[61991]: _type = "Task" [ 871.277709] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.290942] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.324862] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "refresh_cache-97913aea-48c1-4fda-aee3-578e89a08f71" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.325307] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Instance network_info: |[{"id": "3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63", "address": "fa:16:3e:43:c6:46", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c71ee4b-db", "ovs_interfaceid": "3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 871.325777] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:c6:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc845e3-654b-43c6-acea-dde1084f0ad0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 871.333516] env[61991]: DEBUG oslo.service.loopingcall [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.333763] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.334043] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3737d19d-7e23-4e07-a2ee-7e366f202096 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.351173] env[61991]: DEBUG nova.compute.manager [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Received event network-changed-3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.351442] env[61991]: DEBUG nova.compute.manager [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Refreshing instance network info cache due to event network-changed-3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.351835] env[61991]: DEBUG oslo_concurrency.lockutils [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] Acquiring lock "refresh_cache-97913aea-48c1-4fda-aee3-578e89a08f71" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.351945] env[61991]: DEBUG oslo_concurrency.lockutils [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] Acquired lock "refresh_cache-97913aea-48c1-4fda-aee3-578e89a08f71" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.352176] env[61991]: DEBUG nova.network.neutron [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Refreshing network info cache for port 3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.360167] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.360167] env[61991]: value = "task-1129932" [ 871.360167] env[61991]: _type = "Task" [ 871.360167] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.371138] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129932, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.473385] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e6a590-0946-41ac-8057-3087e095f593 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.493808] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778febba-92e0-45bd-be34-98805487634f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.511962] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 871.653637] env[61991]: ERROR nova.scheduler.client.report [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [req-a7ab848b-fad0-4489-9360-2967d7f93e35] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a7ab848b-fad0-4489-9360-2967d7f93e35"}]} [ 871.670458] env[61991]: DEBUG nova.scheduler.client.report [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 871.683889] env[61991]: DEBUG nova.scheduler.client.report [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 871.684205] env[61991]: DEBUG nova.compute.provider_tree [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.696748] env[61991]: DEBUG nova.scheduler.client.report [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 871.714663] env[61991]: DEBUG nova.scheduler.client.report [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 871.797867] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 871.797867] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.798233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.798427] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.798639] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.799944] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-056239ad-718d-4195-b425-f37ae7cf12b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.813752] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.813958] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.814715] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f38bdec4-4336-481e-a252-f07b7b39e13c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.820379] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 871.820379] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52158a15-6c91-5f81-af4f-203ef7cd1b90" [ 871.820379] env[61991]: _type = "Task" [ 871.820379] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.829887] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.830148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.838290] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52158a15-6c91-5f81-af4f-203ef7cd1b90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.853724] env[61991]: DEBUG nova.network.neutron [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updated VIF entry in instance network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.854100] env[61991]: DEBUG nova.network.neutron [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.861457] env[61991]: DEBUG oslo_concurrency.lockutils [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.861726] env[61991]: DEBUG oslo_concurrency.lockutils [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.874979] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129932, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.033837] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84cdcf4-9357-444e-b2d2-697667dca3f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.038517] env[61991]: DEBUG nova.compute.manager [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.038786] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b389eef-d6e5-46b7-8c40-4fac102650ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.049396] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8912e382-05bd-4cdb-8f45-81ac6047a59c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.053039] env[61991]: DEBUG nova.network.neutron [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Port dff90314-9bf7-4448-980c-047fc0a617ec binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 872.084300] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fc120f-7de3-43b1-87db-c6ce4cd8de92 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.092400] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1042226-b1bb-4b87-bd76-09969ccfef08 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.105737] env[61991]: DEBUG nova.compute.provider_tree [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.119769] env[61991]: DEBUG nova.network.neutron [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Updated VIF entry in instance network info cache for port 3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.120155] env[61991]: DEBUG nova.network.neutron [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Updating instance_info_cache with network_info: [{"id": "3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63", "address": "fa:16:3e:43:c6:46", "network": {"id": "dd7aa5a1-b94c-459f-b02c-c283db9d180c", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-624716001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25647cd0715f490092623f4ead833738", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c71ee4b-db", "ovs_interfaceid": "3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.332715] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52158a15-6c91-5f81-af4f-203ef7cd1b90, 'name': SearchDatastore_Task, 'duration_secs': 0.027251} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.333533] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-105ff4c4-499b-4bf7-b23b-2c6a14a8d354 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.338459] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.341295] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 872.341295] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5225b315-650d-2a37-84d5-4956ecf53ecf" [ 872.341295] env[61991]: _type = "Task" [ 872.341295] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.348627] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5225b315-650d-2a37-84d5-4956ecf53ecf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.358621] env[61991]: DEBUG oslo_concurrency.lockutils [req-a5797ec7-87f2-41a8-8034-11955a948ea6 req-cdb238da-a5d3-4261-84ca-b2a4da7c3ebb service nova] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.367520] env[61991]: INFO nova.compute.manager [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Detaching volume 48ca7de5-1564-44be-83d4-3298210a3c5d [ 872.376026] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129932, 'name': CreateVM_Task, 'duration_secs': 0.900053} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.378745] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.379544] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.379713] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.380036] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.380507] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1e5de6f-e173-44b4-bf5a-e51ae0971d3d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.384701] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 872.384701] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524009c1-0df6-a2b7-3571-88923e82fcb7" [ 872.384701] env[61991]: _type = "Task" [ 872.384701] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.392308] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524009c1-0df6-a2b7-3571-88923e82fcb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.407230] env[61991]: INFO nova.virt.block_device [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Attempting to driver detach volume 48ca7de5-1564-44be-83d4-3298210a3c5d from mountpoint /dev/sdb [ 872.407459] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 872.407656] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246849', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'name': 'volume-48ca7de5-1564-44be-83d4-3298210a3c5d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c17f844-1f0e-4f01-aadc-0f1f75a59d06', 'attached_at': '', 'detached_at': '', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'serial': '48ca7de5-1564-44be-83d4-3298210a3c5d'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 872.408455] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1205fce8-6fbe-4fa7-9d45-2628986c9b2e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.428671] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf237c6c-59d4-4b89-9009-b48c4e8d7ee2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.435632] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b523074-e63f-44d9-bab4-df7c37afe5e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.456172] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be573e96-2cfe-40a9-90c5-93dc1143b07e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.470756] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] The volume has not been displaced from its original location: [datastore1] volume-48ca7de5-1564-44be-83d4-3298210a3c5d/volume-48ca7de5-1564-44be-83d4-3298210a3c5d.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 872.476088] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Reconfiguring VM instance instance-0000002a to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 872.476389] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b5615ea-e1bb-46eb-89de-3c93f1343214 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.493806] env[61991]: DEBUG oslo_vmware.api [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 872.493806] env[61991]: value = "task-1129933" [ 872.493806] env[61991]: _type = "Task" [ 872.493806] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.501922] env[61991]: DEBUG oslo_vmware.api [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129933, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.550578] env[61991]: INFO nova.compute.manager [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] instance snapshotting [ 872.553751] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c2d92a-96a3-4f0e-b244-6383323afd39 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.581071] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30c9569-5488-45f1-aaf7-36ebee9f7ef4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.622341] env[61991]: DEBUG oslo_concurrency.lockutils [req-297d2b60-0587-4176-bd9a-024721d0a590 req-b39ac509-cde8-474e-8b05-d08edce50f83 service nova] Releasing lock "refresh_cache-97913aea-48c1-4fda-aee3-578e89a08f71" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.639107] env[61991]: DEBUG nova.scheduler.client.report [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 109 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 872.639404] env[61991]: DEBUG nova.compute.provider_tree [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 109 to 110 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 872.639588] env[61991]: DEBUG nova.compute.provider_tree [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.853958] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5225b315-650d-2a37-84d5-4956ecf53ecf, 'name': SearchDatastore_Task, 'duration_secs': 0.03365} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.854273] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.854537] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. {{(pid=61991) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 872.854789] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe28c6b2-3c45-4908-b1c6-97856c283980 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.861918] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 872.861918] env[61991]: value = "task-1129934" [ 872.861918] env[61991]: _type = "Task" [ 872.861918] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.865874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.870854] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.893076] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524009c1-0df6-a2b7-3571-88923e82fcb7, 'name': SearchDatastore_Task, 'duration_secs': 0.021848} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.893377] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.893608] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.893842] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.893993] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.894188] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.894436] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-602e192b-06f6-4f66-bd6a-569a335f15ea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.902055] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.902242] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.902907] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f12706cb-5535-4f44-b011-966a74c86467 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.907448] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 872.907448] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526d905f-4afc-128f-e043-1da2e8294716" [ 872.907448] env[61991]: _type = "Task" [ 872.907448] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.914794] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526d905f-4afc-128f-e043-1da2e8294716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.002983] env[61991]: DEBUG oslo_vmware.api [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129933, 'name': ReconfigVM_Task, 'duration_secs': 0.230496} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.003309] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Reconfigured VM instance instance-0000002a to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 873.007834] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cf479a3-0e34-48c9-ad4a-4d3f577085be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.021440] env[61991]: DEBUG oslo_vmware.api [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 873.021440] env[61991]: value = "task-1129935" [ 873.021440] env[61991]: _type = "Task" [ 873.021440] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.028790] env[61991]: DEBUG oslo_vmware.api [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.078826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.079034] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.079328] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.092585] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 873.093460] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e8965535-0d0a-433d-9675-300330343908 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.101309] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 873.101309] env[61991]: value = "task-1129936" [ 873.101309] env[61991]: _type = "Task" [ 873.101309] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.109404] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129936, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.145034] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.412s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.145155] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.148247] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.344s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.149684] env[61991]: INFO nova.compute.claims [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.375356] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129934, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.419491] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526d905f-4afc-128f-e043-1da2e8294716, 'name': SearchDatastore_Task, 'duration_secs': 0.031479} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.420428] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11919fa7-ead8-4669-afb7-cbffa90a1395 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.428484] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 873.428484] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287e244-6c09-af91-29fd-bdfff4d0b449" [ 873.428484] env[61991]: _type = "Task" [ 873.428484] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.439992] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287e244-6c09-af91-29fd-bdfff4d0b449, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.531764] env[61991]: DEBUG oslo_vmware.api [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129935, 'name': ReconfigVM_Task, 'duration_secs': 0.132767} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.532171] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246849', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'name': 'volume-48ca7de5-1564-44be-83d4-3298210a3c5d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9c17f844-1f0e-4f01-aadc-0f1f75a59d06', 'attached_at': '', 'detached_at': '', 'volume_id': '48ca7de5-1564-44be-83d4-3298210a3c5d', 'serial': '48ca7de5-1564-44be-83d4-3298210a3c5d'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 873.614056] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129936, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.657061] env[61991]: DEBUG nova.compute.utils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.658124] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.658356] env[61991]: DEBUG nova.network.neutron [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.728347] env[61991]: DEBUG nova.policy [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '160b84053b8a4d7c86b013e751db6b66', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a2b37cc1ade437097583e823d457ca6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.873311] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.731231} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.873628] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. [ 873.874456] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86582f03-70f8-4c4f-b4f3-ee1237faec31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.898886] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.899202] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79c55792-f9c9-4ebe-950b-1ce3736f5073 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.916200] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 873.916200] env[61991]: value = "task-1129937" [ 873.916200] env[61991]: _type = "Task" [ 873.916200] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.923649] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129937, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.937937] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287e244-6c09-af91-29fd-bdfff4d0b449, 'name': SearchDatastore_Task, 'duration_secs': 0.020182} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.938227] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.938488] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 97913aea-48c1-4fda-aee3-578e89a08f71/97913aea-48c1-4fda-aee3-578e89a08f71.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.938739] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ed22f79-458c-45a8-9d73-fd0e6331c202 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.943998] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 873.943998] env[61991]: value = "task-1129938" [ 873.943998] env[61991]: _type = "Task" [ 873.943998] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.951303] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.011719] env[61991]: DEBUG nova.network.neutron [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Successfully created port: 555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.088985] env[61991]: DEBUG nova.objects.instance [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'flavor' on Instance uuid 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.111925] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129936, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.127262] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.127467] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.127674] env[61991]: DEBUG nova.network.neutron [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.163304] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.431689] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129937, 'name': ReconfigVM_Task, 'duration_secs': 0.440184} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.432051] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Reconfigured VM instance instance-0000004e to attach disk [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.432843] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9a2837-f771-4d00-91cd-760b45326b45 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.468693] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3db4c6cb-9f16-4ce3-b647-d6a42d9e863c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.483313] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129938, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.485047] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 874.485047] env[61991]: value = "task-1129939" [ 874.485047] env[61991]: _type = "Task" [ 874.485047] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.494636] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129939, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.525837] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c4b931-c2eb-47f5-b7fb-289a2580ea3b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.533073] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e192e26-2a2f-43eb-bbd1-ea32d1e2bc8a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.563055] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2fbfa6-4ee4-452d-a51f-0f5c6e50437e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.570821] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d66ac99-d5b8-4be7-abca-05bc7e778906 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.585452] env[61991]: DEBUG nova.compute.provider_tree [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.613031] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129936, 'name': CreateSnapshot_Task, 'duration_secs': 1.163492} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.613031] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 874.614351] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee52fbe-a195-43b6-8afb-34bd5496b7ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.942784] env[61991]: DEBUG nova.network.neutron [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.970710] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56755} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.971048] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 97913aea-48c1-4fda-aee3-578e89a08f71/97913aea-48c1-4fda-aee3-578e89a08f71.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.971402] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.971698] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60aa597d-f6d1-4346-8b2c-9c08fa08ecad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.979286] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 874.979286] env[61991]: value = "task-1129940" [ 874.979286] env[61991]: _type = "Task" [ 874.979286] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.987677] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.996445] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129939, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.096374] env[61991]: DEBUG oslo_concurrency.lockutils [None req-211d28bb-0eeb-4fbc-8091-49ddc1048fbf tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.234s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.120233] env[61991]: DEBUG nova.scheduler.client.report [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 110 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 875.120515] env[61991]: DEBUG nova.compute.provider_tree [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 110 to 111 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 875.120704] env[61991]: DEBUG nova.compute.provider_tree [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 875.131956] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 875.132787] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6b4e22d4-b27b-454c-8276-9d007f78c12c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.142435] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 875.142435] env[61991]: value = "task-1129941" [ 875.142435] env[61991]: _type = "Task" [ 875.142435] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.150926] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129941, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.173062] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.197871] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.198159] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.198323] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.198513] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.198664] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.198817] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.199042] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.199214] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.199384] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.199554] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.199732] env[61991]: DEBUG nova.virt.hardware [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.200593] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03611fdb-5113-469d-b048-6084fc19ed17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.208623] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945ca481-0829-4511-a152-9e9b9b9167c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.444439] env[61991]: DEBUG oslo_concurrency.lockutils [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.492712] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.305714} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.497995] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.499664] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c670b14c-fa9f-45b6-a8c8-2c8d13f3602d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.507459] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129939, 'name': ReconfigVM_Task, 'duration_secs': 0.599347} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.516692] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.526472] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 97913aea-48c1-4fda-aee3-578e89a08f71/97913aea-48c1-4fda-aee3-578e89a08f71.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.527903] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4103c8bf-e0f7-4569-ab4f-42df6a32ad9d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.530048] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-895a2605-5e5e-4169-a963-40d64ae8f9e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.546225] env[61991]: DEBUG nova.compute.manager [req-4437c26d-3706-4499-b32d-3fab21422fe7 req-35140b43-73b0-4187-a5ce-13209df0c548 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Received event network-vif-plugged-555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.546521] env[61991]: DEBUG oslo_concurrency.lockutils [req-4437c26d-3706-4499-b32d-3fab21422fe7 req-35140b43-73b0-4187-a5ce-13209df0c548 service nova] Acquiring lock "f902f00f-eb2f-418a-bbfa-535442446681-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.546746] env[61991]: DEBUG oslo_concurrency.lockutils [req-4437c26d-3706-4499-b32d-3fab21422fe7 req-35140b43-73b0-4187-a5ce-13209df0c548 service nova] Lock "f902f00f-eb2f-418a-bbfa-535442446681-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.546920] env[61991]: DEBUG oslo_concurrency.lockutils [req-4437c26d-3706-4499-b32d-3fab21422fe7 req-35140b43-73b0-4187-a5ce-13209df0c548 service nova] Lock "f902f00f-eb2f-418a-bbfa-535442446681-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.547106] env[61991]: DEBUG nova.compute.manager [req-4437c26d-3706-4499-b32d-3fab21422fe7 req-35140b43-73b0-4187-a5ce-13209df0c548 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] No waiting events found dispatching network-vif-plugged-555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.547277] env[61991]: WARNING nova.compute.manager [req-4437c26d-3706-4499-b32d-3fab21422fe7 req-35140b43-73b0-4187-a5ce-13209df0c548 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Received unexpected event network-vif-plugged-555cf7a5-bed5-4484-82ea-cc70ca63465c for instance with vm_state building and task_state spawning. [ 875.555741] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 875.555741] env[61991]: value = "task-1129943" [ 875.555741] env[61991]: _type = "Task" [ 875.555741] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.557179] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 875.557179] env[61991]: value = "task-1129942" [ 875.557179] env[61991]: _type = "Task" [ 875.557179] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.568203] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129943, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.571170] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129942, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.636584] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.636584] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.637481] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.650s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.637747] env[61991]: DEBUG nova.objects.instance [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 875.642281] env[61991]: DEBUG nova.network.neutron [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Successfully updated port: 555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.655668] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129941, 'name': CloneVM_Task} progress is 93%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.971386] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422f481b-975a-4d94-a8a9-73f0cc6566f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.990320] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbe0aac-e945-4ed4-acbb-1a33e42efd6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.997317] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 876.071058] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129943, 'name': ReconfigVM_Task, 'duration_secs': 0.269172} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.074337] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 97913aea-48c1-4fda-aee3-578e89a08f71/97913aea-48c1-4fda-aee3-578e89a08f71.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.074922] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.075158] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e374659-ddfb-472f-bab7-79260f1ac380 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.082768] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 876.082768] env[61991]: value = "task-1129944" [ 876.082768] env[61991]: _type = "Task" [ 876.082768] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.091090] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129944, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.146647] env[61991]: DEBUG nova.compute.utils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.151050] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.151248] env[61991]: DEBUG nova.network.neutron [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.156735] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "refresh_cache-f902f00f-eb2f-418a-bbfa-535442446681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.156880] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "refresh_cache-f902f00f-eb2f-418a-bbfa-535442446681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.157035] env[61991]: DEBUG nova.network.neutron [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.159417] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6adc7e99-589c-4f9d-8628-1725a9e1d789 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.522s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.160303] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.190s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.160529] env[61991]: DEBUG nova.objects.instance [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lazy-loading 'resources' on Instance uuid 96627657-d3d9-4a64-b83e-87323bd10c03 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.174520] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129941, 'name': CloneVM_Task} progress is 93%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.217361] env[61991]: DEBUG nova.policy [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b5d90a913fc41e282e050bb9af628d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adbd4ea9b0744f28a0d4a46b4a04d683', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.265487] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.265794] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.265971] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.266173] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.266522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.268536] env[61991]: INFO nova.compute.manager [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Terminating instance [ 876.270261] env[61991]: DEBUG nova.compute.manager [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.270455] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.271289] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea57ab0-0917-486d-9808-0af565546ab7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.278737] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.278877] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88b8235c-e97f-49d5-b432-a6ca0f7d7da8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.284470] env[61991]: DEBUG oslo_vmware.api [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 876.284470] env[61991]: value = "task-1129945" [ 876.284470] env[61991]: _type = "Task" [ 876.284470] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.292354] env[61991]: DEBUG oslo_vmware.api [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.506200] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-945acfd0-2f86-4e19-a961-848714e77923 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance '3bcde671-5702-4b8a-8881-88eb7dfd0556' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 876.570555] env[61991]: DEBUG nova.network.neutron [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Successfully created port: c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.579366] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.595442] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129944, 'name': Rename_Task, 'duration_secs': 0.142726} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.595660] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.595968] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-460cea51-5d95-4fa4-8dde-79d1dc76671d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.603812] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 876.603812] env[61991]: value = "task-1129946" [ 876.603812] env[61991]: _type = "Task" [ 876.603812] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.613364] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.653133] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.667884] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129941, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.702563] env[61991]: DEBUG nova.network.neutron [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 876.797099] env[61991]: DEBUG oslo_vmware.api [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129945, 'name': PowerOffVM_Task, 'duration_secs': 0.26193} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.797422] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.797598] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.797903] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca42e2d5-bda8-4dc1-9496-a81a6f8abea6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.889634] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.889981] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.891250] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleting the datastore file [datastore1] 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.891250] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb1f620e-3f02-4bf6-a76b-2b1e07e599b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.907257] env[61991]: DEBUG oslo_vmware.api [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 876.907257] env[61991]: value = "task-1129948" [ 876.907257] env[61991]: _type = "Task" [ 876.907257] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.924127] env[61991]: DEBUG oslo_vmware.api [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129948, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.928146] env[61991]: DEBUG nova.network.neutron [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Updating instance_info_cache with network_info: [{"id": "555cf7a5-bed5-4484-82ea-cc70ca63465c", "address": "fa:16:3e:62:97:2c", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap555cf7a5-be", "ovs_interfaceid": "555cf7a5-bed5-4484-82ea-cc70ca63465c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.078844] env[61991]: DEBUG oslo_vmware.api [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129942, 'name': PowerOnVM_Task, 'duration_secs': 1.436606} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.079114] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.082597] env[61991]: DEBUG nova.compute.manager [None req-f7c5b6c1-9a60-4661-b04a-52e11fa5c40e tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.083427] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69b5ba3-ad9e-4a26-b3d9-d1e1a0d21ce4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.087742] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff2b74e-e06d-4729-96ea-414011ee0e7f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.097789] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627b07af-730b-4a91-8232-e90b2d27b17d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.136720] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5575d62a-cf3c-4688-832f-2ba30d4ef0d3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.142854] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129946, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.154855] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34fc35e-332d-4484-8885-2fbae270fe12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.168937] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129941, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.182480] env[61991]: DEBUG nova.compute.provider_tree [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.417620] env[61991]: DEBUG oslo_vmware.api [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1129948, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249686} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.417891] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.418098] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 877.418297] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.418474] env[61991]: INFO nova.compute.manager [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Took 1.15 seconds to destroy the instance on the hypervisor. [ 877.418725] env[61991]: DEBUG oslo.service.loopingcall [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.418909] env[61991]: DEBUG nova.compute.manager [-] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.419016] env[61991]: DEBUG nova.network.neutron [-] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 877.431346] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "refresh_cache-f902f00f-eb2f-418a-bbfa-535442446681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.431668] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Instance network_info: |[{"id": "555cf7a5-bed5-4484-82ea-cc70ca63465c", "address": "fa:16:3e:62:97:2c", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap555cf7a5-be", "ovs_interfaceid": "555cf7a5-bed5-4484-82ea-cc70ca63465c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.432109] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:97:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '555cf7a5-bed5-4484-82ea-cc70ca63465c', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.439611] env[61991]: DEBUG oslo.service.loopingcall [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.439846] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.440130] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a731b3c4-2d1e-4b7f-b8c0-44da04f7753c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.464947] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.464947] env[61991]: value = "task-1129949" [ 877.464947] env[61991]: _type = "Task" [ 877.464947] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.475161] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129949, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.561279] env[61991]: DEBUG nova.compute.manager [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Received event network-changed-555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.561279] env[61991]: DEBUG nova.compute.manager [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Refreshing instance network info cache due to event network-changed-555cf7a5-bed5-4484-82ea-cc70ca63465c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.561700] env[61991]: DEBUG oslo_concurrency.lockutils [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] Acquiring lock "refresh_cache-f902f00f-eb2f-418a-bbfa-535442446681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.561700] env[61991]: DEBUG oslo_concurrency.lockutils [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] Acquired lock "refresh_cache-f902f00f-eb2f-418a-bbfa-535442446681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.561818] env[61991]: DEBUG nova.network.neutron [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Refreshing network info cache for port 555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.624561] env[61991]: DEBUG oslo_vmware.api [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129946, 'name': PowerOnVM_Task, 'duration_secs': 0.59342} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.624871] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.625104] env[61991]: INFO nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Took 8.79 seconds to spawn the instance on the hypervisor. [ 877.625317] env[61991]: DEBUG nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.626204] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c2ed11-40cc-4d58-93c1-addca28ed8c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.660657] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129941, 'name': CloneVM_Task, 'duration_secs': 2.309609} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.660961] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Created linked-clone VM from snapshot [ 877.661817] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a206123-e9d7-4c1f-ba9c-2f3753283526 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.665772] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.678499] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Uploading image 756fdf3f-076f-40eb-b02a-4048ea33d87d {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 877.686445] env[61991]: DEBUG nova.scheduler.client.report [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.701179] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.701639] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.701901] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.702261] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.702715] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.702817] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.703127] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.703392] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.703684] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.703949] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.704233] env[61991]: DEBUG nova.virt.hardware [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.708013] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1b24d8-8a4c-40de-96a7-8e1691192ce0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.723429] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 877.723429] env[61991]: value = "vm-246911" [ 877.723429] env[61991]: _type = "VirtualMachine" [ 877.723429] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 877.725574] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48f6343-9c76-4e1d-b870-8e9622c5c4f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.731781] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-932b1c86-d07e-4cb7-95b5-d11d50b00078 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.748856] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lease: (returnval){ [ 877.748856] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c00d0-ff88-8702-c59f-a1439c4eee1e" [ 877.748856] env[61991]: _type = "HttpNfcLease" [ 877.748856] env[61991]: } obtained for exporting VM: (result){ [ 877.748856] env[61991]: value = "vm-246911" [ 877.748856] env[61991]: _type = "VirtualMachine" [ 877.748856] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 877.750349] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the lease: (returnval){ [ 877.750349] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c00d0-ff88-8702-c59f-a1439c4eee1e" [ 877.750349] env[61991]: _type = "HttpNfcLease" [ 877.750349] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 877.756846] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.756846] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c00d0-ff88-8702-c59f-a1439c4eee1e" [ 877.756846] env[61991]: _type = "HttpNfcLease" [ 877.756846] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 877.826973] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.827300] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.827518] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.827709] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.827886] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.830181] env[61991]: INFO nova.compute.manager [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Terminating instance [ 877.834847] env[61991]: DEBUG nova.compute.manager [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 877.835057] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.836081] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7614125-9ce3-4763-8bb2-ed9f9c1f2cf5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.844209] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.844629] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ecf5fb25-50aa-41ab-bab5-f92f835016fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.851244] env[61991]: DEBUG oslo_vmware.api [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 877.851244] env[61991]: value = "task-1129951" [ 877.851244] env[61991]: _type = "Task" [ 877.851244] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.860323] env[61991]: DEBUG oslo_vmware.api [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.975848] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129949, 'name': CreateVM_Task, 'duration_secs': 0.364887} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.975917] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.977637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.977637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.977637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.978032] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9571c8a2-3ddf-43bc-ac60-e08f10fab990 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.984033] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 877.984033] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521acc53-98be-a31f-9201-80fa3a4c31a4" [ 877.984033] env[61991]: _type = "Task" [ 877.984033] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.993430] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521acc53-98be-a31f-9201-80fa3a4c31a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.001440] env[61991]: DEBUG nova.compute.manager [req-77ce5ee7-4a5f-4b14-a1c7-86eb1c4cf0e2 req-8fea6636-bdb8-4641-8c0d-aaf3fff88df5 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Received event network-vif-deleted-87a47717-3b2c-4160-97b7-ef566ac540c0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.001657] env[61991]: INFO nova.compute.manager [req-77ce5ee7-4a5f-4b14-a1c7-86eb1c4cf0e2 req-8fea6636-bdb8-4641-8c0d-aaf3fff88df5 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Neutron deleted interface 87a47717-3b2c-4160-97b7-ef566ac540c0; detaching it from the instance and deleting it from the info cache [ 878.001981] env[61991]: DEBUG nova.network.neutron [req-77ce5ee7-4a5f-4b14-a1c7-86eb1c4cf0e2 req-8fea6636-bdb8-4641-8c0d-aaf3fff88df5 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.148970] env[61991]: INFO nova.compute.manager [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Took 34.07 seconds to build instance. [ 878.195016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.033s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.199343] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.299s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.199633] env[61991]: DEBUG nova.objects.instance [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lazy-loading 'resources' on Instance uuid c1d4d2eb-810c-42f9-a244-e5aed6e8d517 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.229498] env[61991]: INFO nova.scheduler.client.report [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Deleted allocations for instance 96627657-d3d9-4a64-b83e-87323bd10c03 [ 878.262226] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 878.262226] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c00d0-ff88-8702-c59f-a1439c4eee1e" [ 878.262226] env[61991]: _type = "HttpNfcLease" [ 878.262226] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 878.262226] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 878.262226] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c00d0-ff88-8702-c59f-a1439c4eee1e" [ 878.262226] env[61991]: _type = "HttpNfcLease" [ 878.262226] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 878.262816] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc291731-127d-4d91-8649-2662a86d4713 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.275435] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5293a345-4878-cc45-6db1-684ed7a4114b/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 878.275811] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5293a345-4878-cc45-6db1-684ed7a4114b/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 878.363260] env[61991]: DEBUG oslo_vmware.api [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129951, 'name': PowerOffVM_Task, 'duration_secs': 0.335636} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.363632] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.363980] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.364620] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f639b361-5e6e-4efe-a5a2-451473c582b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.388277] env[61991]: DEBUG nova.network.neutron [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Updated VIF entry in instance network info cache for port 555cf7a5-bed5-4484-82ea-cc70ca63465c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 878.388699] env[61991]: DEBUG nova.network.neutron [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Updating instance_info_cache with network_info: [{"id": "555cf7a5-bed5-4484-82ea-cc70ca63465c", "address": "fa:16:3e:62:97:2c", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap555cf7a5-be", "ovs_interfaceid": "555cf7a5-bed5-4484-82ea-cc70ca63465c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.404607] env[61991]: DEBUG nova.network.neutron [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Successfully updated port: c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.413996] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-243388da-04ea-4774-8b69-b23055b7bceb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.453551] env[61991]: DEBUG nova.network.neutron [-] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.457393] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.457636] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.457843] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleting the datastore file [datastore1] f78ef63b-453e-45d3-959b-4b0c1922b53e {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.458464] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2533ced-eef9-4a45-85be-0520e03545dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.466736] env[61991]: DEBUG oslo_vmware.api [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for the task: (returnval){ [ 878.466736] env[61991]: value = "task-1129953" [ 878.466736] env[61991]: _type = "Task" [ 878.466736] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.476661] env[61991]: DEBUG oslo_vmware.api [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.496141] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521acc53-98be-a31f-9201-80fa3a4c31a4, 'name': SearchDatastore_Task, 'duration_secs': 0.0131} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.496504] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.496764] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.497021] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.497195] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.497838] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.498156] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fce29a57-1a77-4ea9-ac73-8d4806ee8706 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.504664] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f31b4b1c-ed43-4ab5-bb00-b7eb9a76ba28 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.508509] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.508701] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.509964] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51b5dd75-60a6-48b6-8473-a42d1ee2b2c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.518145] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512ce7c5-d977-4b51-84bd-023e5637603c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.532155] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 878.532155] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528dc1ed-d79a-027b-c45b-08c2f20fa4ee" [ 878.532155] env[61991]: _type = "Task" [ 878.532155] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.545593] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528dc1ed-d79a-027b-c45b-08c2f20fa4ee, 'name': SearchDatastore_Task, 'duration_secs': 0.011835} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.557231] env[61991]: DEBUG nova.compute.manager [req-77ce5ee7-4a5f-4b14-a1c7-86eb1c4cf0e2 req-8fea6636-bdb8-4641-8c0d-aaf3fff88df5 service nova] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Detach interface failed, port_id=87a47717-3b2c-4160-97b7-ef566ac540c0, reason: Instance 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 878.557765] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7b88ba-b98b-4013-9b2a-8b57bcc5453e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.564581] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 878.564581] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52089e5f-fe0b-6f82-6a69-5ca58665900d" [ 878.564581] env[61991]: _type = "Task" [ 878.564581] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.574627] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52089e5f-fe0b-6f82-6a69-5ca58665900d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.652671] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc8cb3c0-cd53-41db-a894-bef4a3c391f6 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "97913aea-48c1-4fda-aee3-578e89a08f71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.012s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.746881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3d84c03-f99c-4ecf-952f-1cea9bde62e9 tempest-ServerShowV257Test-1205686756 tempest-ServerShowV257Test-1205686756-project-member] Lock "96627657-d3d9-4a64-b83e-87323bd10c03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.487s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.885660] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "97913aea-48c1-4fda-aee3-578e89a08f71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.886123] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "97913aea-48c1-4fda-aee3-578e89a08f71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.886523] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "97913aea-48c1-4fda-aee3-578e89a08f71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.886949] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "97913aea-48c1-4fda-aee3-578e89a08f71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.887648] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "97913aea-48c1-4fda-aee3-578e89a08f71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.890692] env[61991]: INFO nova.compute.manager [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Terminating instance [ 878.893015] env[61991]: DEBUG oslo_concurrency.lockutils [req-45d4e276-e275-44c8-97af-75b397d6e452 req-e1d3eef3-886a-4975-bfd5-36793c4bcdc4 service nova] Releasing lock "refresh_cache-f902f00f-eb2f-418a-bbfa-535442446681" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.896486] env[61991]: DEBUG nova.compute.manager [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 878.897430] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.898260] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78eeab91-7abd-4a2a-9b10-912ec38c5979 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.908351] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.908923] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.909630] env[61991]: DEBUG nova.network.neutron [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 878.920207] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.921685] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ac8da99-3b9c-4c70-a086-ffc28e4fec24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.933472] env[61991]: DEBUG oslo_vmware.api [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 878.933472] env[61991]: value = "task-1129954" [ 878.933472] env[61991]: _type = "Task" [ 878.933472] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.948562] env[61991]: DEBUG oslo_vmware.api [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.956022] env[61991]: INFO nova.compute.manager [-] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Took 1.54 seconds to deallocate network for instance. [ 878.982541] env[61991]: DEBUG oslo_vmware.api [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Task: {'id': task-1129953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268159} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.982939] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.983106] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.983419] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.983695] env[61991]: INFO nova.compute.manager [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 878.984033] env[61991]: DEBUG oslo.service.loopingcall [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.984300] env[61991]: DEBUG nova.compute.manager [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.984495] env[61991]: DEBUG nova.network.neutron [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 879.077884] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52089e5f-fe0b-6f82-6a69-5ca58665900d, 'name': SearchDatastore_Task, 'duration_secs': 0.011119} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.078862] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.078862] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] f902f00f-eb2f-418a-bbfa-535442446681/f902f00f-eb2f-418a-bbfa-535442446681.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.079260] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-beffda6a-1df1-4718-a6df-e2141928b5c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.081610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.081992] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.082458] env[61991]: DEBUG nova.compute.manager [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Going to confirm migration 2 {{(pid=61991) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 879.098428] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 879.098428] env[61991]: value = "task-1129955" [ 879.098428] env[61991]: _type = "Task" [ 879.098428] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.109840] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.133589] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eceaf8c-4277-4e03-b054-e26ce832fa28 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.144776] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d034d5-431c-4205-b791-d048f379d5d3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.186910] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e62145-92e5-4982-9060-a7387e78fe8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.200463] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdf27cf-7d20-442e-a8cd-cff2a2857fd3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.220529] env[61991]: DEBUG nova.compute.provider_tree [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.450675] env[61991]: DEBUG oslo_vmware.api [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129954, 'name': PowerOffVM_Task, 'duration_secs': 0.315965} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.451374] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.451810] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.451953] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb9e4b11-633d-4792-8783-c3f775b8b66d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.468475] env[61991]: DEBUG nova.network.neutron [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.471459] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.553136] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.553485] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.553694] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleting the datastore file [datastore1] 97913aea-48c1-4fda-aee3-578e89a08f71 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.554022] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-315e5900-dda2-4f14-96c0-a1d591a64b25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.566313] env[61991]: DEBUG oslo_vmware.api [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for the task: (returnval){ [ 879.566313] env[61991]: value = "task-1129957" [ 879.566313] env[61991]: _type = "Task" [ 879.566313] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.578426] env[61991]: DEBUG oslo_vmware.api [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.616881] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129955, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.654142] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.654523] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.654796] env[61991]: DEBUG nova.network.neutron [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.655090] env[61991]: DEBUG nova.objects.instance [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'info_cache' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.719950] env[61991]: DEBUG nova.network.neutron [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating instance_info_cache with network_info: [{"id": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "address": "fa:16:3e:8e:8a:4d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cb6230-aa", "ovs_interfaceid": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.726666] env[61991]: DEBUG nova.scheduler.client.report [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.788848] env[61991]: DEBUG nova.compute.manager [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Received event network-vif-plugged-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.789201] env[61991]: DEBUG oslo_concurrency.lockutils [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.789504] env[61991]: DEBUG oslo_concurrency.lockutils [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.789837] env[61991]: DEBUG oslo_concurrency.lockutils [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.790082] env[61991]: DEBUG nova.compute.manager [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] No waiting events found dispatching network-vif-plugged-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.790295] env[61991]: WARNING nova.compute.manager [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Received unexpected event network-vif-plugged-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 for instance with vm_state building and task_state spawning. [ 879.790932] env[61991]: DEBUG nova.compute.manager [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Received event network-changed-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.790932] env[61991]: DEBUG nova.compute.manager [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Refreshing instance network info cache due to event network-changed-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.791015] env[61991]: DEBUG oslo_concurrency.lockutils [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] Acquiring lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.972941] env[61991]: INFO nova.compute.manager [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Rescuing [ 879.973318] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.973876] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.974154] env[61991]: DEBUG nova.network.neutron [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.076132] env[61991]: DEBUG oslo_vmware.api [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Task: {'id': task-1129957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.384931} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.076556] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.076671] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.076844] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.077063] env[61991]: INFO nova.compute.manager [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Took 1.18 seconds to destroy the instance on the hypervisor. [ 880.077355] env[61991]: DEBUG oslo.service.loopingcall [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.077569] env[61991]: DEBUG nova.compute.manager [-] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.077684] env[61991]: DEBUG nova.network.neutron [-] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.113582] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129955, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628291} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.113795] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] f902f00f-eb2f-418a-bbfa-535442446681/f902f00f-eb2f-418a-bbfa-535442446681.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.114246] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.114699] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6a27309-8d54-4a70-846f-171eef60d202 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.124387] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 880.124387] env[61991]: value = "task-1129958" [ 880.124387] env[61991]: _type = "Task" [ 880.124387] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.136043] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.223083] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.223508] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance network_info: |[{"id": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "address": "fa:16:3e:8e:8a:4d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cb6230-aa", "ovs_interfaceid": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.224034] env[61991]: DEBUG oslo_concurrency.lockutils [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] Acquired lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.224505] env[61991]: DEBUG nova.network.neutron [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Refreshing network info cache for port c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.227033] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:8a:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.235059] env[61991]: DEBUG oslo.service.loopingcall [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.236536] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.038s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.241782] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.242419] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.567s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.242693] env[61991]: DEBUG nova.objects.instance [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lazy-loading 'resources' on Instance uuid 7a468ab7-6ca1-43e2-a9b2-89e7640d8148 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.244360] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35f9465f-9c3e-4f1e-8437-d088e1db5a91 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.262663] env[61991]: DEBUG nova.network.neutron [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.271926] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.271926] env[61991]: value = "task-1129959" [ 880.271926] env[61991]: _type = "Task" [ 880.271926] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.277667] env[61991]: INFO nova.scheduler.client.report [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Deleted allocations for instance c1d4d2eb-810c-42f9-a244-e5aed6e8d517 [ 880.293175] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129959, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.637400] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131479} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.638049] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.639777] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5481c595-e4fa-4e3c-aa7a-d033ea25e390 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.664482] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] f902f00f-eb2f-418a-bbfa-535442446681/f902f00f-eb2f-418a-bbfa-535442446681.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.669620] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-142a974f-ccea-472f-abd8-a4bd8889a33f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.697423] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 880.697423] env[61991]: value = "task-1129960" [ 880.697423] env[61991]: _type = "Task" [ 880.697423] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.708963] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129960, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.764533] env[61991]: INFO nova.compute.manager [-] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Took 1.78 seconds to deallocate network for instance. [ 880.783348] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129959, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.784446] env[61991]: DEBUG nova.network.neutron [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updating instance_info_cache with network_info: [{"id": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "address": "fa:16:3e:14:21:d4", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6f94dd-10", "ovs_interfaceid": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.792330] env[61991]: DEBUG oslo_concurrency.lockutils [None req-615e810e-a914-4a5f-8bc0-0b70bd146d2e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "c1d4d2eb-810c-42f9-a244-e5aed6e8d517" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.861s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.929441] env[61991]: DEBUG nova.network.neutron [-] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.122154] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70eedda9-dea4-42ff-9a82-333d29df66a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.131131] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4ed20a-7d85-4a34-abae-326cf14fa359 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.168697] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8644c18-83e7-4828-8c5e-adace090b6b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.177220] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23df1a52-c233-4384-ab4d-44b62adc08bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.196166] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.196478] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.196696] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.196881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.197066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.199115] env[61991]: DEBUG nova.compute.provider_tree [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.203832] env[61991]: INFO nova.compute.manager [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Terminating instance [ 881.205810] env[61991]: DEBUG nova.compute.manager [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.206015] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.206966] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcde97d-9231-4348-aee5-906841120570 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.212991] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129960, 'name': ReconfigVM_Task, 'duration_secs': 0.41336} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.214459] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Reconfigured VM instance instance-00000053 to attach disk [datastore1] f902f00f-eb2f-418a-bbfa-535442446681/f902f00f-eb2f-418a-bbfa-535442446681.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.215329] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0253b1c5-2530-4560-99d6-fad15978c34e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.219684] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.220249] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26074820-44c0-4759-bf00-be2d5fb8108c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.224724] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 881.224724] env[61991]: value = "task-1129961" [ 881.224724] env[61991]: _type = "Task" [ 881.224724] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.229424] env[61991]: DEBUG oslo_vmware.api [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 881.229424] env[61991]: value = "task-1129962" [ 881.229424] env[61991]: _type = "Task" [ 881.229424] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.236391] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129961, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.241152] env[61991]: DEBUG oslo_vmware.api [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.252862] env[61991]: DEBUG nova.network.neutron [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updated VIF entry in instance network info cache for port c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.253228] env[61991]: DEBUG nova.network.neutron [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating instance_info_cache with network_info: [{"id": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "address": "fa:16:3e:8e:8a:4d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cb6230-aa", "ovs_interfaceid": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.257888] env[61991]: DEBUG nova.network.neutron [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.273470] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.285246] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129959, 'name': CreateVM_Task, 'duration_secs': 0.627387} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.285473] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.286252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.286491] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.286873] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.287460] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.289838] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba505bc-95b7-4bde-a4e2-b2746e234254 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.295235] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 881.295235] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52782511-4df0-178b-0993-aa7c1e841411" [ 881.295235] env[61991]: _type = "Task" [ 881.295235] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.310524] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52782511-4df0-178b-0993-aa7c1e841411, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.433284] env[61991]: INFO nova.compute.manager [-] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Took 1.36 seconds to deallocate network for instance. [ 881.705122] env[61991]: DEBUG nova.scheduler.client.report [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.739799] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129961, 'name': Rename_Task, 'duration_secs': 0.150896} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.742645] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.742919] env[61991]: DEBUG oslo_vmware.api [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129962, 'name': PowerOffVM_Task, 'duration_secs': 0.480473} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.743152] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0629241-edb3-434d-8a87-f45634a17700 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.744601] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.744782] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.745014] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98bf1f28-1398-49aa-b21e-7c0c59e5bb81 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.751660] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 881.751660] env[61991]: value = "task-1129963" [ 881.751660] env[61991]: _type = "Task" [ 881.751660] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.756082] env[61991]: DEBUG oslo_concurrency.lockutils [req-53c72ab8-6976-4843-9bf6-29741f5adc6c req-992c16ad-b7c5-44e7-b10c-4a45378b2f87 service nova] Releasing lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.760038] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.760276] env[61991]: DEBUG nova.objects.instance [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'migration_context' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.761307] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.806840] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52782511-4df0-178b-0993-aa7c1e841411, 'name': SearchDatastore_Task, 'duration_secs': 0.01429} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.807281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.807481] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.807732] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.807890] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.808087] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.808389] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7aa74d8e-c723-4dd6-8468-daf41a8e3965 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.817781] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.818110] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.818274] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Deleting the datastore file [datastore2] f6945e87-3b17-4e0e-845f-3bacf0fbe5ac {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.819525] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-133d5cff-b6a5-4d71-819d-25c6abe25d04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.822408] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.822408] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.824230] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abbfaf58-bced-45b3-9711-ad913bf6e289 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.828404] env[61991]: DEBUG nova.compute.manager [req-2e875a29-839b-465c-b60b-c8f556af6ca2 req-bac2ed74-c3d1-4603-8268-b3149d3f4b2e service nova] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Received event network-vif-deleted-4b255035-ebfa-4e8a-b9e5-a51737f080b9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.828616] env[61991]: DEBUG nova.compute.manager [req-2e875a29-839b-465c-b60b-c8f556af6ca2 req-bac2ed74-c3d1-4603-8268-b3149d3f4b2e service nova] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Received event network-vif-deleted-3c71ee4b-db0a-4d20-8b35-d1cb2ca9bb63 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.832633] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.833533] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b2742d7-705b-4662-8b1e-f8b08d26e472 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.838028] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 881.838028] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528ba450-d0e4-1cc8-7bcd-80a49a6698a8" [ 881.838028] env[61991]: _type = "Task" [ 881.838028] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.838214] env[61991]: DEBUG oslo_vmware.api [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for the task: (returnval){ [ 881.838214] env[61991]: value = "task-1129965" [ 881.838214] env[61991]: _type = "Task" [ 881.838214] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.847421] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 881.847421] env[61991]: value = "task-1129966" [ 881.847421] env[61991]: _type = "Task" [ 881.847421] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.859413] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]528ba450-d0e4-1cc8-7bcd-80a49a6698a8, 'name': SearchDatastore_Task, 'duration_secs': 0.012372} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.859704] env[61991]: DEBUG oslo_vmware.api [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129965, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.861043] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f54d41fb-e64b-4f7c-89ce-d165c67bef99 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.866849] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.870067] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 881.870067] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525980ba-e7ad-d454-6b4a-c6415c2efa79" [ 881.870067] env[61991]: _type = "Task" [ 881.870067] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.878545] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525980ba-e7ad-d454-6b4a-c6415c2efa79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.942265] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.209929] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.967s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.212480] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.383s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.214484] env[61991]: INFO nova.compute.claims [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.238134] env[61991]: INFO nova.scheduler.client.report [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Deleted allocations for instance 7a468ab7-6ca1-43e2-a9b2-89e7640d8148 [ 882.262976] env[61991]: DEBUG nova.objects.base [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Object Instance<3bcde671-5702-4b8a-8881-88eb7dfd0556> lazy-loaded attributes: info_cache,migration_context {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 882.263516] env[61991]: DEBUG oslo_vmware.api [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129963, 'name': PowerOnVM_Task, 'duration_secs': 0.479967} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.264943] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd59d7a-cd8e-4706-b98e-f582b40bc851 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.267917] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.268186] env[61991]: INFO nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Took 7.10 seconds to spawn the instance on the hypervisor. [ 882.268353] env[61991]: DEBUG nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.269385] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b828475f-9395-4b79-b902-502ad9bb7b62 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.292878] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38c336f7-6845-40e9-896b-5270a4f23b65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.301663] env[61991]: DEBUG oslo_vmware.api [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 882.301663] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525f66eb-1001-b1a7-9ea9-dde5c4eb5d9b" [ 882.301663] env[61991]: _type = "Task" [ 882.301663] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.310262] env[61991]: DEBUG oslo_vmware.api [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525f66eb-1001-b1a7-9ea9-dde5c4eb5d9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.349768] env[61991]: DEBUG oslo_vmware.api [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Task: {'id': task-1129965, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.402866} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.353766] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.354016] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.354236] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.354441] env[61991]: INFO nova.compute.manager [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Took 1.15 seconds to destroy the instance on the hypervisor. [ 882.354668] env[61991]: DEBUG oslo.service.loopingcall [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.355706] env[61991]: DEBUG nova.compute.manager [-] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.355780] env[61991]: DEBUG nova.network.neutron [-] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 882.364095] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129966, 'name': PowerOffVM_Task, 'duration_secs': 0.227677} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.364460] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.365374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206d6fe4-0ea0-4930-be4e-cef6dd986c26 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.388499] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9744032e-4896-487f-bb07-dcf2cf9bdd20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.398842] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525980ba-e7ad-d454-6b4a-c6415c2efa79, 'name': SearchDatastore_Task, 'duration_secs': 0.0121} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.401333] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.401610] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.404386] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdeb3dc2-ac74-459f-8395-ad98811f97dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.410370] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 882.410370] env[61991]: value = "task-1129967" [ 882.410370] env[61991]: _type = "Task" [ 882.410370] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.418793] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.426894] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.427247] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e8d85fb-bd13-434f-b424-3f3b61f697bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.435151] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 882.435151] env[61991]: value = "task-1129968" [ 882.435151] env[61991]: _type = "Task" [ 882.435151] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.445196] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 882.445513] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.446260] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.446260] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.446371] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.446643] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5efc286b-5d38-4e87-8e70-db406d9788e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.455654] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.455729] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.456565] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63b5d238-1d2e-40c1-9a04-7d3cac8c3b68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.462458] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 882.462458] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dc7df7-fc76-bd1d-9a9b-e54a4750dd0d" [ 882.462458] env[61991]: _type = "Task" [ 882.462458] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.471374] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dc7df7-fc76-bd1d-9a9b-e54a4750dd0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.625400] env[61991]: DEBUG nova.compute.manager [req-e3074d2d-04b3-49ef-9184-3947eb88985b req-848b6447-2677-4007-8ea9-5bf4e6a7813a service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Received event network-vif-deleted-c9be477d-8647-4664-8ebf-1f1fe5e45492 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.625475] env[61991]: INFO nova.compute.manager [req-e3074d2d-04b3-49ef-9184-3947eb88985b req-848b6447-2677-4007-8ea9-5bf4e6a7813a service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Neutron deleted interface c9be477d-8647-4664-8ebf-1f1fe5e45492; detaching it from the instance and deleting it from the info cache [ 882.625631] env[61991]: DEBUG nova.network.neutron [req-e3074d2d-04b3-49ef-9184-3947eb88985b req-848b6447-2677-4007-8ea9-5bf4e6a7813a service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.746453] env[61991]: DEBUG oslo_concurrency.lockutils [None req-33469e58-02b7-4815-a90b-77319febe225 tempest-ImagesTestJSON-1618990675 tempest-ImagesTestJSON-1618990675-project-member] Lock "7a468ab7-6ca1-43e2-a9b2-89e7640d8148" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.037s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.810025] env[61991]: INFO nova.compute.manager [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Took 31.80 seconds to build instance. [ 882.817056] env[61991]: DEBUG oslo_vmware.api [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525f66eb-1001-b1a7-9ea9-dde5c4eb5d9b, 'name': SearchDatastore_Task, 'duration_secs': 0.009136} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.817056] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.921726] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129967, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.974084] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dc7df7-fc76-bd1d-9a9b-e54a4750dd0d, 'name': SearchDatastore_Task, 'duration_secs': 0.011966} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.975097] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d44feb7b-6c6a-40e7-9f78-42483dd76d62 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.982487] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 882.982487] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52033069-68f2-cd71-dc98-b8a630710108" [ 882.982487] env[61991]: _type = "Task" [ 882.982487] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.991086] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52033069-68f2-cd71-dc98-b8a630710108, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.097120] env[61991]: DEBUG nova.network.neutron [-] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.130072] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f25e0e45-1225-4cc0-b26a-14bb80058012 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.140860] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2b3690-a8ce-4c55-b61b-dfd6c3abdef4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.180051] env[61991]: DEBUG nova.compute.manager [req-e3074d2d-04b3-49ef-9184-3947eb88985b req-848b6447-2677-4007-8ea9-5bf4e6a7813a service nova] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Detach interface failed, port_id=c9be477d-8647-4664-8ebf-1f1fe5e45492, reason: Instance f6945e87-3b17-4e0e-845f-3bacf0fbe5ac could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 883.313057] env[61991]: DEBUG oslo_concurrency.lockutils [None req-66d1987a-6a98-4480-a0e1-cbca4a3cfac8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "f902f00f-eb2f-418a-bbfa-535442446681" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.319s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.425270] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670705} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.425688] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.425981] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.429056] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-528b9051-35d8-4708-9c84-b30e73ba9cce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.437385] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 883.437385] env[61991]: value = "task-1129969" [ 883.437385] env[61991]: _type = "Task" [ 883.437385] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.449360] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.492768] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52033069-68f2-cd71-dc98-b8a630710108, 'name': SearchDatastore_Task, 'duration_secs': 0.022838} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.495397] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.495684] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. {{(pid=61991) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 883.496145] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df03ebd2-b14f-436e-8093-98de2679192a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.503530] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 883.503530] env[61991]: value = "task-1129970" [ 883.503530] env[61991]: _type = "Task" [ 883.503530] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.514826] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.555218] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbc189f-e938-4c02-8878-8320993e3db2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.561908] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a494b1ef-bc63-4cc6-abcb-f68b13d0f0fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.591793] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fa9c10-8270-4cdc-aab9-304eca3a1dce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.598635] env[61991]: INFO nova.compute.manager [-] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Took 1.24 seconds to deallocate network for instance. [ 883.601401] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c60b027-3c9d-428a-800f-8233eb18bc24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.620011] env[61991]: DEBUG nova.compute.provider_tree [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.660826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "f902f00f-eb2f-418a-bbfa-535442446681" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.661107] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "f902f00f-eb2f-418a-bbfa-535442446681" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.661319] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "f902f00f-eb2f-418a-bbfa-535442446681-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.661518] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "f902f00f-eb2f-418a-bbfa-535442446681-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.661705] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "f902f00f-eb2f-418a-bbfa-535442446681-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.663709] env[61991]: INFO nova.compute.manager [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Terminating instance [ 883.665932] env[61991]: DEBUG nova.compute.manager [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.666141] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.667070] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecec558-5db2-4e7b-9256-c07a4a974a2c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.674367] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.674597] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edf306a2-ccf9-4a36-9de9-e71dfa4b1285 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.680886] env[61991]: DEBUG oslo_vmware.api [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 883.680886] env[61991]: value = "task-1129971" [ 883.680886] env[61991]: _type = "Task" [ 883.680886] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.691568] env[61991]: DEBUG oslo_vmware.api [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129971, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.949426] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.17185} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.949802] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.950794] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5487f4-21fa-4509-840c-a1d6ec0c7da9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.974612] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.974854] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a2b7299-b9cf-4112-8e8b-32bb8cb5480e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.995408] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 883.995408] env[61991]: value = "task-1129972" [ 883.995408] env[61991]: _type = "Task" [ 883.995408] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.007226] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129972, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.017492] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129970, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.109965] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.123483] env[61991]: DEBUG nova.scheduler.client.report [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.194684] env[61991]: DEBUG oslo_vmware.api [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129971, 'name': PowerOffVM_Task, 'duration_secs': 0.229973} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.195039] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.195252] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.198404] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9dc34782-6739-4906-bbe4-a20d0b0ca40a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.269973] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.269973] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.269973] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleting the datastore file [datastore1] f902f00f-eb2f-418a-bbfa-535442446681 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.269973] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5cd22b29-77b1-4a0e-ac4f-3589b586c099 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.276447] env[61991]: DEBUG oslo_vmware.api [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 884.276447] env[61991]: value = "task-1129974" [ 884.276447] env[61991]: _type = "Task" [ 884.276447] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.285266] env[61991]: DEBUG oslo_vmware.api [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129974, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.506254] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.515731] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.756389} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.516262] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. [ 884.516896] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19d73ad-b70b-4470-8a1e-1d06bc30ab4d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.546766] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.547127] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaeaf0e2-bc78-4679-8d80-33e061131d14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.572167] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 884.572167] env[61991]: value = "task-1129975" [ 884.572167] env[61991]: _type = "Task" [ 884.572167] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.580932] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129975, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.629026] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.629646] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.632715] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.716s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.632921] env[61991]: DEBUG nova.objects.instance [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 3183190e-3202-40be-894c-e47b186708b5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 884.788862] env[61991]: DEBUG oslo_vmware.api [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1129974, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342524} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.789180] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.789373] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.789547] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.789721] env[61991]: INFO nova.compute.manager [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Took 1.12 seconds to destroy the instance on the hypervisor. [ 884.789961] env[61991]: DEBUG oslo.service.loopingcall [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.790165] env[61991]: DEBUG nova.compute.manager [-] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.790258] env[61991]: DEBUG nova.network.neutron [-] [instance: f902f00f-eb2f-418a-bbfa-535442446681] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 885.006956] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129972, 'name': ReconfigVM_Task, 'duration_secs': 0.661976} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.007326] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfigured VM instance instance-00000054 to attach disk [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.008126] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f9cc80a-e789-4edb-b9cf-f929a7933d4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.015583] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 885.015583] env[61991]: value = "task-1129976" [ 885.015583] env[61991]: _type = "Task" [ 885.015583] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.023867] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129976, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.084760] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129975, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.095856] env[61991]: DEBUG nova.compute.manager [req-c8694ef4-f62f-4bde-87fb-b46198b76738 req-078ee317-18e5-46e1-a9ba-30afc32706e0 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Received event network-vif-deleted-555cf7a5-bed5-4484-82ea-cc70ca63465c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.096120] env[61991]: INFO nova.compute.manager [req-c8694ef4-f62f-4bde-87fb-b46198b76738 req-078ee317-18e5-46e1-a9ba-30afc32706e0 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Neutron deleted interface 555cf7a5-bed5-4484-82ea-cc70ca63465c; detaching it from the instance and deleting it from the info cache [ 885.097355] env[61991]: DEBUG nova.network.neutron [req-c8694ef4-f62f-4bde-87fb-b46198b76738 req-078ee317-18e5-46e1-a9ba-30afc32706e0 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.137766] env[61991]: DEBUG nova.compute.utils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.141852] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.142051] env[61991]: DEBUG nova.network.neutron [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 885.187603] env[61991]: DEBUG nova.policy [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e6810d3aced4a44bbfe9ba49270bc35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a18b2a73f7643e1b89c9660a770e9da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.497803] env[61991]: DEBUG nova.network.neutron [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Successfully created port: 5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.528132] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129976, 'name': Rename_Task, 'duration_secs': 0.324138} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.528132] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.528909] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c56fc0d4-2b97-49d9-8703-c461ecb074d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.535389] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 885.535389] env[61991]: value = "task-1129977" [ 885.535389] env[61991]: _type = "Task" [ 885.535389] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.543692] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129977, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.573539] env[61991]: DEBUG nova.network.neutron [-] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.583059] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129975, 'name': ReconfigVM_Task, 'duration_secs': 0.583486} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.583722] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.584363] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386fcb46-183b-4b05-bf96-8dabf15dcc4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.617444] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b07ff099-5737-4ab8-986a-ecae5fdb4f34 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.620007] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-317b3f3f-fad6-413f-bb77-6039dd6dabcf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.638385] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0b78d4-b464-4e92-ba75-b11ad4b81d85 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.652943] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.657108] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc9fb189-1fef-418e-bc95-1da57f4e93aa tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.024s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.658178] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 885.658178] env[61991]: value = "task-1129978" [ 885.658178] env[61991]: _type = "Task" [ 885.658178] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.658392] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.760s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.670512] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.683458] env[61991]: DEBUG nova.compute.manager [req-c8694ef4-f62f-4bde-87fb-b46198b76738 req-078ee317-18e5-46e1-a9ba-30afc32706e0 service nova] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Detach interface failed, port_id=555cf7a5-bed5-4484-82ea-cc70ca63465c, reason: Instance f902f00f-eb2f-418a-bbfa-535442446681 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 886.051859] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129977, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.075230] env[61991]: INFO nova.compute.manager [-] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Took 1.28 seconds to deallocate network for instance. [ 886.178705] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129978, 'name': ReconfigVM_Task, 'duration_secs': 0.210754} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.181027] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.181027] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65dbd145-9730-4aa3-9d83-2a0943c89eb5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.187189] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 886.187189] env[61991]: value = "task-1129979" [ 886.187189] env[61991]: _type = "Task" [ 886.187189] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.197154] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129979, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.547438] env[61991]: DEBUG oslo_vmware.api [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1129977, 'name': PowerOnVM_Task, 'duration_secs': 0.636619} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.547753] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.547965] env[61991]: INFO nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Took 8.88 seconds to spawn the instance on the hypervisor. [ 886.548162] env[61991]: DEBUG nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.548934] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b795f4c-8b1c-40cf-89b7-8df0f180e44b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.582568] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.674422] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.676973] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Applying migration context for instance 3bcde671-5702-4b8a-8881-88eb7dfd0556 as it has an incoming, in-progress migration f8d56f1f-b98b-44cc-a80d-b069ddd2017b. Migration status is confirming {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 886.678555] env[61991]: INFO nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating resource usage from migration f8d56f1f-b98b-44cc-a80d-b069ddd2017b [ 886.701316] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129979, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.703774] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.704087] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.704246] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.704453] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.704639] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.704806] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.705068] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.705457] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.705752] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.706487] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.706487] env[61991]: DEBUG nova.virt.hardware [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.707519] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.707519] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.707681] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f6945e87-3b17-4e0e-845f-3bacf0fbe5ac is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.707718] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d36bd981-9ea3-46f6-8376-ac1e0c3bf61e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.707848] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f78ef63b-453e-45d3-959b-4b0c1922b53e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.708387] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 56188232-bed0-4a4b-a4bc-01edbb85cbe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.708387] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3183190e-3202-40be-894c-e47b186708b5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.708387] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.708387] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f7eab1a9-2a8a-4664-8736-ca599e207c31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.708572] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 7634e00f-48ee-40a0-8625-2f8dc83cd58d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.708613] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9d1ac179-3d7e-4f0b-a80d-033eba447cab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.708730] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.708866] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 97913aea-48c1-4fda-aee3-578e89a08f71 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 886.708991] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Migration f8d56f1f-b98b-44cc-a80d-b069ddd2017b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 886.709124] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3bcde671-5702-4b8a-8881-88eb7dfd0556 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.709263] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f902f00f-eb2f-418a-bbfa-535442446681 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.709378] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance b3a2ea40-5c4c-4e7d-95b1-38a18c429cec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.709549] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 39004ce4-c96b-4911-9b64-3c508ddcecaf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 886.711685] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8999d613-f41b-4ea2-af6e-7b6c304a688d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.724238] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c5c20d-4ac3-4ef1-942b-847348d8d5fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.072924] env[61991]: INFO nova.compute.manager [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Took 35.29 seconds to build instance. [ 887.124745] env[61991]: DEBUG nova.compute.manager [req-1de12cab-1d49-40d1-a15d-bb151b2488d1 req-761fe67a-5f65-460d-bd50-4f7d8fd0b8cb service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Received event network-vif-plugged-5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.124745] env[61991]: DEBUG oslo_concurrency.lockutils [req-1de12cab-1d49-40d1-a15d-bb151b2488d1 req-761fe67a-5f65-460d-bd50-4f7d8fd0b8cb service nova] Acquiring lock "39004ce4-c96b-4911-9b64-3c508ddcecaf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.128690] env[61991]: DEBUG oslo_concurrency.lockutils [req-1de12cab-1d49-40d1-a15d-bb151b2488d1 req-761fe67a-5f65-460d-bd50-4f7d8fd0b8cb service nova] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.128690] env[61991]: DEBUG oslo_concurrency.lockutils [req-1de12cab-1d49-40d1-a15d-bb151b2488d1 req-761fe67a-5f65-460d-bd50-4f7d8fd0b8cb service nova] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.128690] env[61991]: DEBUG nova.compute.manager [req-1de12cab-1d49-40d1-a15d-bb151b2488d1 req-761fe67a-5f65-460d-bd50-4f7d8fd0b8cb service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] No waiting events found dispatching network-vif-plugged-5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.128690] env[61991]: WARNING nova.compute.manager [req-1de12cab-1d49-40d1-a15d-bb151b2488d1 req-761fe67a-5f65-460d-bd50-4f7d8fd0b8cb service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Received unexpected event network-vif-plugged-5346674f-dcb2-496e-9948-ef1c9c184532 for instance with vm_state building and task_state spawning. [ 887.199604] env[61991]: DEBUG oslo_vmware.api [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129979, 'name': PowerOnVM_Task, 'duration_secs': 0.73222} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.199961] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.202820] env[61991]: DEBUG nova.compute.manager [None req-cf3ef57e-5771-4488-a34b-95c2050005fa tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.203736] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c55a52-8c67-4902-8350-da390b1b679c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.223626] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 0dcf415e-b554-4c82-b127-9f0e22176031 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 887.238675] env[61991]: DEBUG nova.network.neutron [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Successfully updated port: 5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.578809] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a5914b07-4f57-4294-bec8-df437edc0aa5 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.802s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.683808] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.684323] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.732202] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 902f158c-fcc4-40a2-abbd-491bdfd4b77a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 887.733309] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 887.733992] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 887.743408] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-39004ce4-c96b-4911-9b64-3c508ddcecaf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.743408] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-39004ce4-c96b-4911-9b64-3c508ddcecaf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.743494] env[61991]: DEBUG nova.network.neutron [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.970606] env[61991]: DEBUG nova.compute.manager [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Received event network-changed-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.971185] env[61991]: DEBUG nova.compute.manager [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Refreshing instance network info cache due to event network-changed-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.971185] env[61991]: DEBUG oslo_concurrency.lockutils [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] Acquiring lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.971185] env[61991]: DEBUG oslo_concurrency.lockutils [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] Acquired lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.971429] env[61991]: DEBUG nova.network.neutron [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Refreshing network info cache for port c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.089161] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a81ba2c-925a-435c-999d-d7d5106e73de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.097337] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e24d1c-465f-4872-af5e-366ffd4e8cee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.138779] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e7394b-7f5f-44a4-9c0a-1e72adc65d91 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.141823] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.142067] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.148539] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0658a22f-cfd6-4168-b656-ad765b37d2bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.163050] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.187021] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.287094] env[61991]: DEBUG nova.network.neutron [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.433164] env[61991]: DEBUG nova.network.neutron [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Updating instance_info_cache with network_info: [{"id": "5346674f-dcb2-496e-9948-ef1c9c184532", "address": "fa:16:3e:6b:96:82", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5346674f-dc", "ovs_interfaceid": "5346674f-dcb2-496e-9948-ef1c9c184532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.568740] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "077c5def-84cc-48f9-be90-410100666773" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.568971] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "077c5def-84cc-48f9-be90-410100666773" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.644373] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.666411] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.680702] env[61991]: INFO nova.compute.manager [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Unrescuing [ 888.682048] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.682048] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.682048] env[61991]: DEBUG nova.network.neutron [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.719755] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.760663] env[61991]: DEBUG nova.network.neutron [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updated VIF entry in instance network info cache for port c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.760663] env[61991]: DEBUG nova.network.neutron [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating instance_info_cache with network_info: [{"id": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "address": "fa:16:3e:8e:8a:4d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2cb6230-aa", "ovs_interfaceid": "c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.937069] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-39004ce4-c96b-4911-9b64-3c508ddcecaf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.937553] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Instance network_info: |[{"id": "5346674f-dcb2-496e-9948-ef1c9c184532", "address": "fa:16:3e:6b:96:82", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5346674f-dc", "ovs_interfaceid": "5346674f-dcb2-496e-9948-ef1c9c184532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.938115] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:96:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5346674f-dcb2-496e-9948-ef1c9c184532', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.949506] env[61991]: DEBUG oslo.service.loopingcall [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.949875] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.950269] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39c2e316-a2fd-4f76-a4e2-afd1f022b0a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.978925] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.978925] env[61991]: value = "task-1129980" [ 888.978925] env[61991]: _type = "Task" [ 888.978925] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.987420] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129980, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.074910] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.155687] env[61991]: DEBUG nova.compute.manager [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Received event network-changed-5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.155941] env[61991]: DEBUG nova.compute.manager [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Refreshing instance network info cache due to event network-changed-5346674f-dcb2-496e-9948-ef1c9c184532. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.156177] env[61991]: DEBUG oslo_concurrency.lockutils [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] Acquiring lock "refresh_cache-39004ce4-c96b-4911-9b64-3c508ddcecaf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.156325] env[61991]: DEBUG oslo_concurrency.lockutils [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] Acquired lock "refresh_cache-39004ce4-c96b-4911-9b64-3c508ddcecaf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.156515] env[61991]: DEBUG nova.network.neutron [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Refreshing network info cache for port 5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 889.169677] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.170667] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 889.170922] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.513s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.171222] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.649s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.172737] env[61991]: INFO nova.compute.claims [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.262850] env[61991]: DEBUG oslo_concurrency.lockutils [req-298058c4-9289-4502-955b-fbbbfe75bd17 req-dd866189-cd49-470d-88a1-f1d85e6a1759 service nova] Releasing lock "refresh_cache-b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.453027] env[61991]: DEBUG nova.network.neutron [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updating instance_info_cache with network_info: [{"id": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "address": "fa:16:3e:14:21:d4", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb6f94dd-10", "ovs_interfaceid": "db6f94dd-1023-4cf6-8997-5bead8bfa1b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.490433] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129980, 'name': CreateVM_Task, 'duration_secs': 0.340743} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.490633] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 889.491449] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.491664] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.492445] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 889.493030] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94d5aa2c-ac69-40f3-9296-b0d5e203f0e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.499039] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 889.499039] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5220a324-96b5-db3c-255f-9af4978f844f" [ 889.499039] env[61991]: _type = "Task" [ 889.499039] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.508062] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5220a324-96b5-db3c-255f-9af4978f844f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.594489] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.886762] env[61991]: DEBUG nova.network.neutron [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Updated VIF entry in instance network info cache for port 5346674f-dcb2-496e-9948-ef1c9c184532. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.887212] env[61991]: DEBUG nova.network.neutron [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Updating instance_info_cache with network_info: [{"id": "5346674f-dcb2-496e-9948-ef1c9c184532", "address": "fa:16:3e:6b:96:82", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5346674f-dc", "ovs_interfaceid": "5346674f-dcb2-496e-9948-ef1c9c184532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.954963] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.955701] env[61991]: DEBUG nova.objects.instance [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'flavor' on Instance uuid 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.010057] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5220a324-96b5-db3c-255f-9af4978f844f, 'name': SearchDatastore_Task, 'duration_secs': 0.015302} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.011025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.011025] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.011025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.011025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.011286] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.011453] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d38404c1-3900-48e9-89b5-25e8cb55fdfa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.026963] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.027178] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.027901] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-640f6bc6-9189-46c0-abca-1b2cc7d380f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.033172] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 890.033172] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522d97a9-96a7-d00a-ab1e-7dd07a6bea61" [ 890.033172] env[61991]: _type = "Task" [ 890.033172] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.042714] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522d97a9-96a7-d00a-ab1e-7dd07a6bea61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.390471] env[61991]: DEBUG oslo_concurrency.lockutils [req-827c0e48-b6af-445f-b5a9-e5abdb56e024 req-00559dd3-b063-405f-85b2-86ad71a8f482 service nova] Releasing lock "refresh_cache-39004ce4-c96b-4911-9b64-3c508ddcecaf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.463970] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87edc0b2-c2e5-4f1d-bd03-46da1de23b5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.485274] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.487325] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-beb568d3-4979-44f9-bc67-f8a305b4c0ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.490599] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59bb95b-77f7-44ee-9001-e66b600145c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.498548] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c41ee9-ac74-43fd-8c33-96c5c3759274 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.501668] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 890.501668] env[61991]: value = "task-1129981" [ 890.501668] env[61991]: _type = "Task" [ 890.501668] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.531422] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6340cff-dca5-420f-927c-1ba4c577a630 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.536996] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.542668] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5293a345-4878-cc45-6db1-684ed7a4114b/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 890.543452] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4dbd664-b2c5-451a-89f0-daf0997d09a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.550683] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522d97a9-96a7-d00a-ab1e-7dd07a6bea61, 'name': SearchDatastore_Task, 'duration_secs': 0.03677} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.552658] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb09bc4d-620e-43a3-891e-b86537d41a16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.557982] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-678cf5f2-5fa0-40db-b943-3f267b0e2eaf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.559930] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5293a345-4878-cc45-6db1-684ed7a4114b/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 890.560187] env[61991]: ERROR oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5293a345-4878-cc45-6db1-684ed7a4114b/disk-0.vmdk due to incomplete transfer. [ 890.560715] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b6cf65c4-a38f-460f-84cb-42107f66aa9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.571583] env[61991]: DEBUG nova.compute.provider_tree [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.573831] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 890.573831] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5292f273-3a87-75d6-7fa2-a1827e1c7de9" [ 890.573831] env[61991]: _type = "Task" [ 890.573831] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.586021] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5292f273-3a87-75d6-7fa2-a1827e1c7de9, 'name': SearchDatastore_Task, 'duration_secs': 0.010694} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.586021] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.586021] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 39004ce4-c96b-4911-9b64-3c508ddcecaf/39004ce4-c96b-4911-9b64-3c508ddcecaf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.586021] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15baa504-b1ba-4cbe-aad9-0b250499637b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.586277] env[61991]: DEBUG oslo_vmware.rw_handles [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5293a345-4878-cc45-6db1-684ed7a4114b/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 890.586277] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Uploaded image 756fdf3f-076f-40eb-b02a-4048ea33d87d to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 890.588728] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 890.589816] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9ccf4277-a52b-4c96-bf6e-df01c07caede {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.594519] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 890.594519] env[61991]: value = "task-1129982" [ 890.594519] env[61991]: _type = "Task" [ 890.594519] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.597824] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 890.597824] env[61991]: value = "task-1129983" [ 890.597824] env[61991]: _type = "Task" [ 890.597824] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.604385] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.608985] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129983, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.011595] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129981, 'name': PowerOffVM_Task, 'duration_secs': 0.224756} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.011884] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.017417] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.017784] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-325eae78-51b9-41a9-8c5e-2f12127b5b6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.037431] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 891.037431] env[61991]: value = "task-1129984" [ 891.037431] env[61991]: _type = "Task" [ 891.037431] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.045602] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129984, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.075803] env[61991]: DEBUG nova.scheduler.client.report [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.107031] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129982, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456547} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.110107] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 39004ce4-c96b-4911-9b64-3c508ddcecaf/39004ce4-c96b-4911-9b64-3c508ddcecaf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.110395] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.110688] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129983, 'name': Destroy_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.110924] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbef88a1-a710-4fe8-92d4-33b1dda7f7a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.117436] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 891.117436] env[61991]: value = "task-1129985" [ 891.117436] env[61991]: _type = "Task" [ 891.117436] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.124511] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129985, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.547999] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129984, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.583030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.583030] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.585332] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.847s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.585597] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.587625] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.722s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.589108] env[61991]: INFO nova.compute.claims [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.608212] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129983, 'name': Destroy_Task, 'duration_secs': 0.544214} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.608468] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Destroyed the VM [ 891.608716] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 891.608978] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d655c39c-d520-4940-807c-a330ad49e786 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.615824] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 891.615824] env[61991]: value = "task-1129986" [ 891.615824] env[61991]: _type = "Task" [ 891.615824] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.627820] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129985, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068148} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.629155] env[61991]: INFO nova.scheduler.client.report [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocations for instance 3183190e-3202-40be-894c-e47b186708b5 [ 891.632880] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.633775] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129986, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.638412] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e50d66-e316-476a-8767-dd5d0fd7cf1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.660752] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 39004ce4-c96b-4911-9b64-3c508ddcecaf/39004ce4-c96b-4911-9b64-3c508ddcecaf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.661317] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c799b5b-41e3-49ff-9c7b-9ce4ccd2c303 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.681032] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 891.681032] env[61991]: value = "task-1129987" [ 891.681032] env[61991]: _type = "Task" [ 891.681032] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.692260] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.048091] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129984, 'name': ReconfigVM_Task, 'duration_secs': 0.545242} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.048091] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 892.048244] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.049027] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3887e7a-4dfc-4882-910c-9f3fa98618d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.054784] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 892.054784] env[61991]: value = "task-1129988" [ 892.054784] env[61991]: _type = "Task" [ 892.054784] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.062759] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129988, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.094781] env[61991]: DEBUG nova.compute.utils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.099251] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.099251] env[61991]: DEBUG nova.network.neutron [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.126126] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129986, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.143968] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6b318e4-f875-4d3c-ab2e-106ba19eba10 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "3183190e-3202-40be-894c-e47b186708b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.368s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.146260] env[61991]: DEBUG nova.policy [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6422dfa4c424fbaab78c3e72a0e6669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dddb51029854105bc6c9c8724181d39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.191260] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.505327] env[61991]: DEBUG nova.network.neutron [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Successfully created port: 4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.566928] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129988, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.599923] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.639681] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129986, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.695712] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129987, 'name': ReconfigVM_Task, 'duration_secs': 0.836336} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.696017] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 39004ce4-c96b-4911-9b64-3c508ddcecaf/39004ce4-c96b-4911-9b64-3c508ddcecaf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.696717] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f76c922-d428-4f12-8338-00a0dfe06f81 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.704310] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 892.704310] env[61991]: value = "task-1129989" [ 892.704310] env[61991]: _type = "Task" [ 892.704310] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.718349] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129989, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.949072] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16dbec31-4bbb-4301-8ca4-c1c0e602ead8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.956572] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cff5191-8bde-431f-811a-77d359ceb3f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.987856] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0ebb66-fbea-4d33-a36e-72a1d6c14b42 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.995348] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80d9ac0-fa7b-4f6d-bf35-325997ba07f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.008848] env[61991]: DEBUG nova.compute.provider_tree [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.064802] env[61991]: DEBUG oslo_vmware.api [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1129988, 'name': PowerOnVM_Task, 'duration_secs': 0.661008} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.065092] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.065362] env[61991]: DEBUG nova.compute.manager [None req-e98dd60f-a7c8-4ae0-88ea-5dd628b3d065 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.066145] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a85eaa-f318-49fe-8b72-8703b530b92d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.129290] env[61991]: DEBUG oslo_vmware.api [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129986, 'name': RemoveSnapshot_Task, 'duration_secs': 1.268351} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.129601] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 893.129845] env[61991]: INFO nova.compute.manager [None req-14204237-d15f-4682-9932-6858624cb419 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Took 20.58 seconds to snapshot the instance on the hypervisor. [ 893.214344] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129989, 'name': Rename_Task, 'duration_secs': 0.165048} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.214619] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.214865] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d05927f-e1f1-47a8-a443-03ca1a8a156c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.220911] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 893.220911] env[61991]: value = "task-1129990" [ 893.220911] env[61991]: _type = "Task" [ 893.220911] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.228336] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129990, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.512165] env[61991]: DEBUG nova.scheduler.client.report [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.611599] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.639833] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.640110] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.640272] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.640455] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.640636] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.640751] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.640953] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.641128] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.641296] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.641459] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.641635] env[61991]: DEBUG nova.virt.hardware [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.642511] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f4792a-f578-4c24-b70f-4180d2fd866d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.651764] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88625cd0-534e-46df-b96e-2d235e2f124e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.733390] env[61991]: DEBUG oslo_vmware.api [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129990, 'name': PowerOnVM_Task, 'duration_secs': 0.455389} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.733565] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.733858] env[61991]: INFO nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Took 7.06 seconds to spawn the instance on the hypervisor. [ 893.734186] env[61991]: DEBUG nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.735228] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e87c63-cf92-4e08-8ae8-28980c7b452e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.017231] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.017768] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.026891] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.554s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.026891] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.027276] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.754s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.027510] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.029285] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.087s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.031326] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.033145] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 11.217s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.071774] env[61991]: INFO nova.scheduler.client.report [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Deleted allocations for instance 97913aea-48c1-4fda-aee3-578e89a08f71 [ 894.078766] env[61991]: INFO nova.scheduler.client.report [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted allocations for instance 9c17f844-1f0e-4f01-aadc-0f1f75a59d06 [ 894.094070] env[61991]: INFO nova.scheduler.client.report [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Deleted allocations for instance f78ef63b-453e-45d3-959b-4b0c1922b53e [ 894.102336] env[61991]: DEBUG nova.compute.manager [req-cc35f43d-36fe-4441-a613-b1b34342e094 req-5993c55e-f570-4298-b70d-4efecded33c8 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Received event network-vif-plugged-4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.103024] env[61991]: DEBUG oslo_concurrency.lockutils [req-cc35f43d-36fe-4441-a613-b1b34342e094 req-5993c55e-f570-4298-b70d-4efecded33c8 service nova] Acquiring lock "0dcf415e-b554-4c82-b127-9f0e22176031-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.103224] env[61991]: DEBUG oslo_concurrency.lockutils [req-cc35f43d-36fe-4441-a613-b1b34342e094 req-5993c55e-f570-4298-b70d-4efecded33c8 service nova] Lock "0dcf415e-b554-4c82-b127-9f0e22176031-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.105231] env[61991]: DEBUG oslo_concurrency.lockutils [req-cc35f43d-36fe-4441-a613-b1b34342e094 req-5993c55e-f570-4298-b70d-4efecded33c8 service nova] Lock "0dcf415e-b554-4c82-b127-9f0e22176031-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.105231] env[61991]: DEBUG nova.compute.manager [req-cc35f43d-36fe-4441-a613-b1b34342e094 req-5993c55e-f570-4298-b70d-4efecded33c8 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] No waiting events found dispatching network-vif-plugged-4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.105231] env[61991]: WARNING nova.compute.manager [req-cc35f43d-36fe-4441-a613-b1b34342e094 req-5993c55e-f570-4298-b70d-4efecded33c8 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Received unexpected event network-vif-plugged-4b76d924-ef23-455b-afc4-7ea48006e69e for instance with vm_state building and task_state spawning. [ 894.210771] env[61991]: DEBUG nova.network.neutron [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Successfully updated port: 4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.255565] env[61991]: INFO nova.compute.manager [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Took 29.45 seconds to build instance. [ 894.536993] env[61991]: DEBUG nova.compute.utils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.543220] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.543469] env[61991]: DEBUG nova.network.neutron [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 894.581254] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8b2e34a2-6ae9-4ffd-b38a-b66a59b41529 tempest-ImagesOneServerNegativeTestJSON-1267542044 tempest-ImagesOneServerNegativeTestJSON-1267542044-project-member] Lock "97913aea-48c1-4fda-aee3-578e89a08f71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.695s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.596484] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0003f476-0cac-47d8-997e-c887d5201873 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "9c17f844-1f0e-4f01-aadc-0f1f75a59d06" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.331s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.604179] env[61991]: DEBUG nova.policy [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.612512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2ca74f10-abaa-4824-bd24-652befd874db tempest-AttachVolumeShelveTestJSON-1482537930 tempest-AttachVolumeShelveTestJSON-1482537930-project-member] Lock "f78ef63b-453e-45d3-959b-4b0c1922b53e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.785s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.716485] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-0dcf415e-b554-4c82-b127-9f0e22176031" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.716866] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-0dcf415e-b554-4c82-b127-9f0e22176031" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.716866] env[61991]: DEBUG nova.network.neutron [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 894.760336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257c0ecc-bc4a-428a-b346-2fc29eadeb77 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.961s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.868012] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732531a2-030e-4019-a6bd-744547302021 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.877574] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460a5610-8d45-4150-903a-9cd732ef645a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.908338] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81811abd-a31d-43ba-9325-248fa89b2519 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.915927] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bcde63-d1f4-4783-afb7-7991b5b8c840 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.930016] env[61991]: DEBUG nova.compute.provider_tree [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.047128] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.092750] env[61991]: DEBUG nova.network.neutron [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Successfully created port: e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 895.184891] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab9bdb6-ff65-4dcc-8d78-42ca074f3769 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.195332] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Suspending the VM {{(pid=61991) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 895.195593] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6401c902-ccf2-4487-a46d-22d4bc15ef72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.206792] env[61991]: DEBUG oslo_vmware.api [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 895.206792] env[61991]: value = "task-1129991" [ 895.206792] env[61991]: _type = "Task" [ 895.206792] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.217065] env[61991]: DEBUG oslo_vmware.api [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129991, 'name': SuspendVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.287870] env[61991]: DEBUG nova.network.neutron [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.434478] env[61991]: DEBUG nova.scheduler.client.report [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.600986] env[61991]: DEBUG nova.compute.manager [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.602938] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edcfe29b-b4ce-4fc9-a136-b465c142471c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.662191] env[61991]: DEBUG nova.network.neutron [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Updating instance_info_cache with network_info: [{"id": "4b76d924-ef23-455b-afc4-7ea48006e69e", "address": "fa:16:3e:a7:1e:c1", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b76d924-ef", "ovs_interfaceid": "4b76d924-ef23-455b-afc4-7ea48006e69e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.718480] env[61991]: DEBUG oslo_vmware.api [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129991, 'name': SuspendVM_Task} progress is 62%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.061139] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.087102] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.087102] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.087102] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.087102] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.087102] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.087472] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.087894] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.088262] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.088629] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.089090] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.089432] env[61991]: DEBUG nova.virt.hardware [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.090773] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9098d8fe-245b-47fb-825c-b40bf2ddf675 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.100218] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1f66ec-82ad-4fab-87b3-bcbf4f32c859 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.119971] env[61991]: INFO nova.compute.manager [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] instance snapshotting [ 896.124084] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b43b58-0035-4e78-a471-9c73c2ae707c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.144567] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6216cc-2158-44dc-8e81-3d0443422a37 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.165045] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-0dcf415e-b554-4c82-b127-9f0e22176031" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.165473] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance network_info: |[{"id": "4b76d924-ef23-455b-afc4-7ea48006e69e", "address": "fa:16:3e:a7:1e:c1", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b76d924-ef", "ovs_interfaceid": "4b76d924-ef23-455b-afc4-7ea48006e69e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.165906] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:1e:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b76d924-ef23-455b-afc4-7ea48006e69e', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.173539] env[61991]: DEBUG oslo.service.loopingcall [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.173815] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.174090] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c82202ec-9c3e-4528-96e3-5846bee0863a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.196829] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.196829] env[61991]: value = "task-1129992" [ 896.196829] env[61991]: _type = "Task" [ 896.196829] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.216138] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129992, 'name': CreateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.226041] env[61991]: DEBUG oslo_vmware.api [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129991, 'name': SuspendVM_Task, 'duration_secs': 0.698935} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.226041] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Suspended the VM {{(pid=61991) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 896.226041] env[61991]: DEBUG nova.compute.manager [None req-3205cfc1-f1a2-465a-832b-8c35c36fe4ad tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.226740] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9253c62d-46f3-4d4f-97a4-30b86e6b597c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.300144] env[61991]: DEBUG nova.compute.manager [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Received event network-changed-4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.300396] env[61991]: DEBUG nova.compute.manager [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Refreshing instance network info cache due to event network-changed-4b76d924-ef23-455b-afc4-7ea48006e69e. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.300644] env[61991]: DEBUG oslo_concurrency.lockutils [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] Acquiring lock "refresh_cache-0dcf415e-b554-4c82-b127-9f0e22176031" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.300766] env[61991]: DEBUG oslo_concurrency.lockutils [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] Acquired lock "refresh_cache-0dcf415e-b554-4c82-b127-9f0e22176031" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.300930] env[61991]: DEBUG nova.network.neutron [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Refreshing network info cache for port 4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.448131] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.413s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.448131] env[61991]: DEBUG nova.compute.manager [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61991) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 896.449064] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.339s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.449279] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.455350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.873s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.455599] env[61991]: DEBUG nova.objects.instance [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'resources' on Instance uuid f902f00f-eb2f-418a-bbfa-535442446681 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.492808] env[61991]: INFO nova.scheduler.client.report [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Deleted allocations for instance f6945e87-3b17-4e0e-845f-3bacf0fbe5ac [ 896.655180] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 896.655601] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1cd88a3f-a588-4c32-8ab7-92b6e3c9d695 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.664928] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 896.664928] env[61991]: value = "task-1129993" [ 896.664928] env[61991]: _type = "Task" [ 896.664928] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.673729] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129993, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.707454] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129992, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.873439] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.873769] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.874931] env[61991]: INFO nova.compute.manager [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Rebooting instance [ 897.000026] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ac7017d2-642e-409a-9c6e-0f8e72e75c3e tempest-ServerRescueTestJSON-967733375 tempest-ServerRescueTestJSON-967733375-project-member] Lock "f6945e87-3b17-4e0e-845f-3bacf0fbe5ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.803s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.032627] env[61991]: INFO nova.scheduler.client.report [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted allocation for migration f8d56f1f-b98b-44cc-a80d-b069ddd2017b [ 897.179223] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129993, 'name': CreateSnapshot_Task, 'duration_secs': 0.483854} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.179223] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 897.180172] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df637e1d-faf7-4967-989c-d30c04d2828d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.209987] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129992, 'name': CreateVM_Task, 'duration_secs': 0.534691} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.213285] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.214981] env[61991]: DEBUG nova.network.neutron [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Updated VIF entry in instance network info cache for port 4b76d924-ef23-455b-afc4-7ea48006e69e. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.214981] env[61991]: DEBUG nova.network.neutron [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Updating instance_info_cache with network_info: [{"id": "4b76d924-ef23-455b-afc4-7ea48006e69e", "address": "fa:16:3e:a7:1e:c1", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b76d924-ef", "ovs_interfaceid": "4b76d924-ef23-455b-afc4-7ea48006e69e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.216906] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.217050] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.217363] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.223320] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c036bec3-ef77-4eab-8d0c-e94c11be28f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.223320] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 897.223320] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c390b4-975b-755f-bb19-568652946ee8" [ 897.223320] env[61991]: _type = "Task" [ 897.223320] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.234356] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c390b4-975b-755f-bb19-568652946ee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.279245] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f850d1ba-ee17-49e8-aebb-e1c4201f365a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.288329] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ed487e-5409-40b0-bc0f-15142716cbcb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.322283] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa24345c-a403-404d-b895-ada40feae3f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.329907] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a415ed6-4e45-43f6-b165-db8a72950208 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.347184] env[61991]: DEBUG nova.compute.provider_tree [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.352028] env[61991]: DEBUG nova.network.neutron [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Successfully updated port: e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.392945] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.393161] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.393340] env[61991]: DEBUG nova.network.neutron [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 897.538474] env[61991]: DEBUG oslo_concurrency.lockutils [None req-47cda972-7444-4f55-93d9-ba4b57cb82da tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 18.456s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.675526] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.675526] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.699345] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 897.699345] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-93838e7f-00b6-409c-ad6c-9bc9cb259881 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.709926] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 897.709926] env[61991]: value = "task-1129994" [ 897.709926] env[61991]: _type = "Task" [ 897.709926] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.721161] env[61991]: DEBUG oslo_concurrency.lockutils [req-eb5aac22-4b65-402c-8aeb-73b09bcd8046 req-648402d8-7bd9-4dd5-82f0-a648952af819 service nova] Releasing lock "refresh_cache-0dcf415e-b554-4c82-b127-9f0e22176031" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.721710] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129994, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.736519] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c390b4-975b-755f-bb19-568652946ee8, 'name': SearchDatastore_Task, 'duration_secs': 0.054127} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.737146] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.738175] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.738479] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.738590] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.738781] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.739117] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c2dec43-9711-46e3-86eb-59075f288e61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.753535] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.753535] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.753535] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f8643fc-add7-4a1d-9923-75b0e8ea8d32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.758744] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 897.758744] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52004ab2-4e60-c8f1-a79a-8180a247f2c7" [ 897.758744] env[61991]: _type = "Task" [ 897.758744] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.768488] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52004ab2-4e60-c8f1-a79a-8180a247f2c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.851832] env[61991]: DEBUG nova.scheduler.client.report [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.857265] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.857265] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.857265] env[61991]: DEBUG nova.network.neutron [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.178998] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.222391] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129994, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.249252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "39004ce4-c96b-4911-9b64-3c508ddcecaf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.249252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.249252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "39004ce4-c96b-4911-9b64-3c508ddcecaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.249252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.249252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.249252] env[61991]: INFO nova.compute.manager [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Terminating instance [ 898.251481] env[61991]: DEBUG nova.compute.manager [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.251676] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.252978] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a498cd15-c605-4bf8-97cd-626c0e6655d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.266416] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.267168] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ed5d3c8-fc70-4e7e-a4d7-f77ed53d79fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.275884] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52004ab2-4e60-c8f1-a79a-8180a247f2c7, 'name': SearchDatastore_Task, 'duration_secs': 0.019038} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.277284] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d0abc4a-e716-41cc-b6c2-45b70b104d3e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.283934] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 898.283934] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287ad4c-44aa-86f4-d773-982ad3afb1d8" [ 898.283934] env[61991]: _type = "Task" [ 898.283934] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.295136] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287ad4c-44aa-86f4-d773-982ad3afb1d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.328314] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.328314] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.337958] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.337958] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.337958] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleting the datastore file [datastore2] 39004ce4-c96b-4911-9b64-3c508ddcecaf {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.338551] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0afb77f2-f2ec-42a4-b653-09d90dd8b1b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.346088] env[61991]: DEBUG oslo_vmware.api [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 898.346088] env[61991]: value = "task-1129996" [ 898.346088] env[61991]: _type = "Task" [ 898.346088] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.355555] env[61991]: DEBUG oslo_vmware.api [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.360384] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.364576] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.645s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.366083] env[61991]: INFO nova.compute.claims [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.393390] env[61991]: INFO nova.scheduler.client.report [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted allocations for instance f902f00f-eb2f-418a-bbfa-535442446681 [ 898.459451] env[61991]: DEBUG nova.network.neutron [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 898.470790] env[61991]: DEBUG nova.compute.manager [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-vif-plugged-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.471089] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.471343] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.471571] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.471780] env[61991]: DEBUG nova.compute.manager [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] No waiting events found dispatching network-vif-plugged-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.471983] env[61991]: WARNING nova.compute.manager [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received unexpected event network-vif-plugged-e0abb70f-8b4c-4450-a784-ec0dd2381264 for instance with vm_state building and task_state spawning. [ 898.472377] env[61991]: DEBUG nova.compute.manager [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.472577] env[61991]: DEBUG nova.compute.manager [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing instance network info cache due to event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.472992] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.507491] env[61991]: DEBUG nova.network.neutron [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.698235] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.720206] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129994, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.736422] env[61991]: DEBUG nova.network.neutron [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.795097] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5287ad4c-44aa-86f4-d773-982ad3afb1d8, 'name': SearchDatastore_Task, 'duration_secs': 0.013051} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.795313] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.795644] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.795905] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a53aedb7-fae4-44c4-b3ad-d2c8ac5306d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.802836] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 898.802836] env[61991]: value = "task-1129997" [ 898.802836] env[61991]: _type = "Task" [ 898.802836] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.811958] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.830354] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 898.856363] env[61991]: DEBUG oslo_vmware.api [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1129996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262175} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.859822] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.859822] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.859822] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.859822] env[61991]: INFO nova.compute.manager [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Took 0.61 seconds to destroy the instance on the hypervisor. [ 898.859822] env[61991]: DEBUG oslo.service.loopingcall [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.859822] env[61991]: DEBUG nova.compute.manager [-] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.859822] env[61991]: DEBUG nova.network.neutron [-] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 898.865567] env[61991]: DEBUG nova.objects.instance [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'flavor' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.906294] env[61991]: DEBUG oslo_concurrency.lockutils [None req-adba9099-bc76-40a1-87fa-3f8f8d530ec7 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "f902f00f-eb2f-418a-bbfa-535442446681" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.245s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.010413] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.015999] env[61991]: DEBUG nova.compute.manager [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.019518] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db58ddc-cbf2-4ede-a514-ac0ce1e577e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.222657] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1129994, 'name': CloneVM_Task, 'duration_secs': 1.353641} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.222720] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Created linked-clone VM from snapshot [ 899.223736] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a57be7-4073-4b36-a38f-d0af7ea81365 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.235016] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Uploading image bb20df08-8e30-49f4-988e-4841fc317709 {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 899.240781] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.240781] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Instance network_info: |[{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.240781] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.240946] env[61991]: DEBUG nova.network.neutron [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 899.242316] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:10:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3f695b6-65bc-45cc-a61d-3c38a14e5c0c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0abb70f-8b4c-4450-a784-ec0dd2381264', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.251747] env[61991]: DEBUG oslo.service.loopingcall [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.258182] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.258182] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b82234a-0a6b-4f9f-821a-3e78d4f907c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.288213] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.288213] env[61991]: value = "task-1129998" [ 899.288213] env[61991]: _type = "Task" [ 899.288213] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.298650] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 899.298650] env[61991]: value = "vm-246917" [ 899.298650] env[61991]: _type = "VirtualMachine" [ 899.298650] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 899.298650] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f68a069b-4f30-407e-aee3-754139ac377a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.301955] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129998, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.310267] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lease: (returnval){ [ 899.310267] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a82ba0-89d6-90c2-e716-bb425f330087" [ 899.310267] env[61991]: _type = "HttpNfcLease" [ 899.310267] env[61991]: } obtained for exporting VM: (result){ [ 899.310267] env[61991]: value = "vm-246917" [ 899.310267] env[61991]: _type = "VirtualMachine" [ 899.310267] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 899.311332] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the lease: (returnval){ [ 899.311332] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a82ba0-89d6-90c2-e716-bb425f330087" [ 899.311332] env[61991]: _type = "HttpNfcLease" [ 899.311332] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 899.320048] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129997, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.327061] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.327061] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a82ba0-89d6-90c2-e716-bb425f330087" [ 899.327061] env[61991]: _type = "HttpNfcLease" [ 899.327061] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 899.327061] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 899.327061] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a82ba0-89d6-90c2-e716-bb425f330087" [ 899.327061] env[61991]: _type = "HttpNfcLease" [ 899.327061] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 899.328233] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493735c3-d860-4faf-94aa-ed736477c720 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.341765] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e58007-796b-142f-7380-afd983384c6b/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 899.341991] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e58007-796b-142f-7380-afd983384c6b/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 899.417637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.418224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.418391] env[61991]: DEBUG oslo_concurrency.lockutils [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.418565] env[61991]: DEBUG nova.network.neutron [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.418747] env[61991]: DEBUG nova.objects.instance [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'info_cache' on Instance uuid 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.451743] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5c3507e5-f33e-46c8-870a-75bb2e9b3b41 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.742834] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03445ab-ec5b-47fc-9a32-c66e70941c2e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.751717] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc0300b-eda6-475f-b229-4446dfc71cc8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.794493] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac8f360-d01e-444d-9ee4-eb4315ec56f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.803216] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129998, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.811355] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca0225d-3f55-465b-998e-3ce30b19bd52 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.818746] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1129997, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612002} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.830120] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 899.830642] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.830907] env[61991]: DEBUG nova.network.neutron [-] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.832412] env[61991]: DEBUG nova.compute.provider_tree [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.836970] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa5e370f-8692-4314-a934-6f06fb59587c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.842336] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 899.842336] env[61991]: value = "task-1130000" [ 899.842336] env[61991]: _type = "Task" [ 899.842336] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.854105] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130000, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.925390] env[61991]: DEBUG nova.objects.base [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Object Instance<3bcde671-5702-4b8a-8881-88eb7dfd0556> lazy-loaded attributes: flavor,info_cache {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 900.042747] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b44d12-3742-4ee9-8aff-adadb5ecac9b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.059107] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Doing hard reboot of VM {{(pid=61991) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 900.059471] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-cc641887-9f22-4b03-8bfb-16fa7eb9c63c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.066999] env[61991]: DEBUG oslo_vmware.api [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 900.066999] env[61991]: value = "task-1130001" [ 900.066999] env[61991]: _type = "Task" [ 900.066999] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.079765] env[61991]: DEBUG oslo_vmware.api [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130001, 'name': ResetVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.155021] env[61991]: DEBUG nova.network.neutron [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updated VIF entry in instance network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 900.155021] env[61991]: DEBUG nova.network.neutron [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.318160] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1129998, 'name': CreateVM_Task, 'duration_secs': 0.655454} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.318160] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.318160] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.318160] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.318160] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.318160] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f769d112-2cd5-4280-85b2-6c21e98f5539 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.325489] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 900.325489] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5274c772-5973-8110-e201-ae857e446641" [ 900.325489] env[61991]: _type = "Task" [ 900.325489] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.334885] env[61991]: INFO nova.compute.manager [-] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Took 1.48 seconds to deallocate network for instance. [ 900.354983] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5274c772-5973-8110-e201-ae857e446641, 'name': SearchDatastore_Task, 'duration_secs': 0.014063} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.360042] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.360042] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.360259] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.360483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.360748] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.361122] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16f694dc-4949-41e5-b224-82a2d950522f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.369451] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130000, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076807} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.370162] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.371261] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcc3359-f919-49d4-8d90-68d90b6d2697 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.378347] env[61991]: ERROR nova.scheduler.client.report [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [req-a3a16592-46f5-4765-99c9-e8884f6a71ba] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a3a16592-46f5-4765-99c9-e8884f6a71ba"}]} [ 900.385116] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.385327] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.387167] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da4d4fc1-1764-4f21-aafd-be883be234fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.408461] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.409773] env[61991]: DEBUG nova.scheduler.client.report [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 900.412846] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b2a8d3b-17d1-4bcd-a813-18bacbce68c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.436737] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 900.436737] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ec0812-d9c4-54d3-a02b-bf5180d9e2f1" [ 900.436737] env[61991]: _type = "Task" [ 900.436737] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.445644] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 900.445644] env[61991]: value = "task-1130002" [ 900.445644] env[61991]: _type = "Task" [ 900.445644] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.450754] env[61991]: DEBUG nova.scheduler.client.report [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 900.451072] env[61991]: DEBUG nova.compute.provider_tree [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.457687] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ec0812-d9c4-54d3-a02b-bf5180d9e2f1, 'name': SearchDatastore_Task, 'duration_secs': 0.021333} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.459498] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83897289-881e-4866-a674-a5c71921aa5b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.465137] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130002, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.469057] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 900.469057] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529e9ad4-6f44-d745-921e-d5b7f414ea8f" [ 900.469057] env[61991]: _type = "Task" [ 900.469057] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.474261] env[61991]: DEBUG nova.scheduler.client.report [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 900.482148] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529e9ad4-6f44-d745-921e-d5b7f414ea8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.498467] env[61991]: DEBUG nova.scheduler.client.report [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 900.512861] env[61991]: DEBUG nova.compute.manager [req-ef8fdfd1-94dc-4d06-8561-6b6601ea4ce9 req-b6d52df5-58ce-4891-a15e-71a1f0886f5f service nova] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Received event network-vif-deleted-5346674f-dcb2-496e-9948-ef1c9c184532 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.577413] env[61991]: DEBUG oslo_vmware.api [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130001, 'name': ResetVM_Task, 'duration_secs': 0.094709} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.577818] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Did hard reboot of VM {{(pid=61991) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 900.578087] env[61991]: DEBUG nova.compute.manager [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.579321] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3cc09f-f4ca-4454-91ee-14afeb37409d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.655954] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0e4538f-66a7-4477-8319-04493602f287 req-e76ac688-59b9-4d45-8872-7b1bd9fa2cfe service nova] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.787203] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90e9c31-0bf3-4a59-93f5-72fadf8ad8ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.795093] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a57395-9d9a-4842-ab57-6560acf9511f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.830707] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ca07bf-adf4-4490-87b2-2a026571521e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.838690] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74f4cd8-9be7-4178-896d-8f87a3406464 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.853479] env[61991]: DEBUG nova.compute.provider_tree [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.858148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.957332] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130002, 'name': ReconfigVM_Task, 'duration_secs': 0.494258} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.957708] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.958471] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12ded4f1-d67e-4554-96cd-ea24e271080a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.965417] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 900.965417] env[61991]: value = "task-1130003" [ 900.965417] env[61991]: _type = "Task" [ 900.965417] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.978564] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130003, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.984991] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529e9ad4-6f44-d745-921e-d5b7f414ea8f, 'name': SearchDatastore_Task, 'duration_secs': 0.012533} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.985291] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.985598] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 902f158c-fcc4-40a2-abbd-491bdfd4b77a/902f158c-fcc4-40a2-abbd-491bdfd4b77a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 900.986204] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7c6160d-fdde-4b28-9e65-07d9070cd705 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.992853] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 900.992853] env[61991]: value = "task-1130004" [ 900.992853] env[61991]: _type = "Task" [ 900.992853] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.002091] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.100515] env[61991]: DEBUG oslo_concurrency.lockutils [None req-32f7d336-e2fc-4878-a810-9ff72cc45328 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.224s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.143693] env[61991]: DEBUG nova.network.neutron [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [{"id": "dff90314-9bf7-4448-980c-047fc0a617ec", "address": "fa:16:3e:81:2d:26", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdff90314-9b", "ovs_interfaceid": "dff90314-9bf7-4448-980c-047fc0a617ec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.396346] env[61991]: DEBUG nova.scheduler.client.report [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 901.396766] env[61991]: DEBUG nova.compute.provider_tree [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 117 to 118 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 901.397036] env[61991]: DEBUG nova.compute.provider_tree [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.480169] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130003, 'name': Rename_Task, 'duration_secs': 0.221942} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.480623] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.480950] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdba8580-00e8-43c1-9925-11ee299a8615 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.490076] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 901.490076] env[61991]: value = "task-1130005" [ 901.490076] env[61991]: _type = "Task" [ 901.490076] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.505328] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130004, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.510082] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130005, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.647750] env[61991]: DEBUG oslo_concurrency.lockutils [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-3bcde671-5702-4b8a-8881-88eb7dfd0556" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.904082] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.540s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.904823] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 901.911816] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.742s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.916693] env[61991]: INFO nova.compute.claims [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.007535] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130005, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.012224] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130004, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.740067} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.013034] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 902f158c-fcc4-40a2-abbd-491bdfd4b77a/902f158c-fcc4-40a2-abbd-491bdfd4b77a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 902.013196] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.013576] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdf5d2b7-fa3f-450e-b23f-0b012d3142d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.021448] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 902.021448] env[61991]: value = "task-1130006" [ 902.021448] env[61991]: _type = "Task" [ 902.021448] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.030734] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130006, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.150943] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.151391] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bce62cb6-d312-46e5-8f8c-9f5494b2a949 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.158957] env[61991]: DEBUG oslo_vmware.api [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 902.158957] env[61991]: value = "task-1130007" [ 902.158957] env[61991]: _type = "Task" [ 902.158957] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.167788] env[61991]: DEBUG oslo_vmware.api [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130007, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.413589] env[61991]: DEBUG nova.compute.utils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.415812] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 902.416140] env[61991]: DEBUG nova.network.neutron [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 902.458081] env[61991]: DEBUG nova.policy [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be32a956d6f6471faabf7a1861800290', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c696231bd014ee5a67e4c3719f94ab4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.508662] env[61991]: DEBUG oslo_vmware.api [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130005, 'name': PowerOnVM_Task, 'duration_secs': 0.558514} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.508978] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.509312] env[61991]: INFO nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Took 8.90 seconds to spawn the instance on the hypervisor. [ 902.509603] env[61991]: DEBUG nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.512102] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe7e314-7c95-4386-b7be-39de9a4407c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.531068] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094054} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.531619] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.533330] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79b793d-d810-41b6-8484-218aa6a17e69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.560534] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 902f158c-fcc4-40a2-abbd-491bdfd4b77a/902f158c-fcc4-40a2-abbd-491bdfd4b77a.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.562923] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bf04ec4-a2a2-4fc3-9153-e03f2470e1c9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.584099] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 902.584099] env[61991]: value = "task-1130008" [ 902.584099] env[61991]: _type = "Task" [ 902.584099] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.593065] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.669434] env[61991]: DEBUG oslo_vmware.api [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130007, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.722192] env[61991]: DEBUG nova.network.neutron [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Successfully created port: 9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.919961] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.034727] env[61991]: INFO nova.compute.manager [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Took 34.53 seconds to build instance. [ 903.102986] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.173545] env[61991]: DEBUG oslo_vmware.api [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130007, 'name': PowerOnVM_Task, 'duration_secs': 0.653032} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.173869] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.174092] env[61991]: DEBUG nova.compute.manager [None req-53f45ad6-e29a-4b2d-aec1-08e2872e2b9c tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.174980] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75fe134-24bd-45f0-abe9-75ead6b18280 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.232509] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6eb8751-426c-4dc6-bd9d-55914525d9c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.240408] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c17434-5e82-40c8-9dd9-f9d4f321cff8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.270744] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b323a44-59a0-4489-b06a-0781fb8d91a4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.278217] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc90b20a-5920-4222-a32a-f01166dc92d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.291138] env[61991]: DEBUG nova.compute.provider_tree [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.536951] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec2e8f44-6ace-4192-a3be-43b4ff53744f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "0dcf415e-b554-4c82-b127-9f0e22176031" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.043s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.598833] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130008, 'name': ReconfigVM_Task, 'duration_secs': 0.821681} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.599247] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 902f158c-fcc4-40a2-abbd-491bdfd4b77a/902f158c-fcc4-40a2-abbd-491bdfd4b77a.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.599951] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a84a6126-2958-4c6e-b117-ee70745437d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.606827] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 903.606827] env[61991]: value = "task-1130009" [ 903.606827] env[61991]: _type = "Task" [ 903.606827] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.615581] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130009, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.797224] env[61991]: DEBUG nova.scheduler.client.report [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.933757] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 903.967158] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.967781] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.968074] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.968346] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.968640] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.968874] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.969145] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.969362] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.969587] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.969798] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.970030] env[61991]: DEBUG nova.virt.hardware [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.971157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fafed4-8f02-4d1c-bf97-511e825ad7cf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.980512] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b6a9cc8-5058-4f0d-8017-51abc798c91c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.024819] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.025305] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.117212] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130009, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.147387] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.147387] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.147546] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.151033] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.151033] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.151033] env[61991]: INFO nova.compute.manager [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Terminating instance [ 904.152205] env[61991]: DEBUG nova.compute.manager [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.152402] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.153331] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f671ef-6f04-4696-8145-61233e10bc59 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.161357] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.161591] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96603894-f6b5-4e82-993e-2530c4b91bf4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.167711] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 904.167711] env[61991]: value = "task-1130010" [ 904.167711] env[61991]: _type = "Task" [ 904.167711] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.176810] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.300938] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.301545] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.304586] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.710s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.306343] env[61991]: INFO nova.compute.claims [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.406251] env[61991]: DEBUG nova.compute.manager [req-d4785462-bf3e-4a72-9b0e-6dfb9db624ca req-218dd748-cf72-4512-a0e0-b8bad7b7a1bc service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Received event network-vif-plugged-9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.406486] env[61991]: DEBUG oslo_concurrency.lockutils [req-d4785462-bf3e-4a72-9b0e-6dfb9db624ca req-218dd748-cf72-4512-a0e0-b8bad7b7a1bc service nova] Acquiring lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.408483] env[61991]: DEBUG oslo_concurrency.lockutils [req-d4785462-bf3e-4a72-9b0e-6dfb9db624ca req-218dd748-cf72-4512-a0e0-b8bad7b7a1bc service nova] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.408483] env[61991]: DEBUG oslo_concurrency.lockutils [req-d4785462-bf3e-4a72-9b0e-6dfb9db624ca req-218dd748-cf72-4512-a0e0-b8bad7b7a1bc service nova] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.408483] env[61991]: DEBUG nova.compute.manager [req-d4785462-bf3e-4a72-9b0e-6dfb9db624ca req-218dd748-cf72-4512-a0e0-b8bad7b7a1bc service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] No waiting events found dispatching network-vif-plugged-9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.408483] env[61991]: WARNING nova.compute.manager [req-d4785462-bf3e-4a72-9b0e-6dfb9db624ca req-218dd748-cf72-4512-a0e0-b8bad7b7a1bc service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Received unexpected event network-vif-plugged-9b9cd766-2c3d-4f0a-beb3-30531d95c2ff for instance with vm_state building and task_state spawning. [ 904.492525] env[61991]: DEBUG nova.network.neutron [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Successfully updated port: 9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.531851] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 904.620177] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130009, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.678966] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.810576] env[61991]: DEBUG nova.compute.utils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.814639] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.814907] env[61991]: DEBUG nova.network.neutron [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 904.870773] env[61991]: DEBUG nova.policy [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be32a956d6f6471faabf7a1861800290', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c696231bd014ee5a67e4c3719f94ab4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.906346] env[61991]: INFO nova.compute.manager [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Rebuilding instance [ 904.965412] env[61991]: DEBUG nova.compute.manager [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.966441] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d2cb3c-cc49-4210-aabf-ebc40527b398 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.996433] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.997270] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.997270] env[61991]: DEBUG nova.network.neutron [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.054507] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.124853] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130009, 'name': Rename_Task, 'duration_secs': 1.173402} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.125702] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.125702] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-680b4b9a-b177-44bd-bfb4-93057de88852 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.133040] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 905.133040] env[61991]: value = "task-1130011" [ 905.133040] env[61991]: _type = "Task" [ 905.133040] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.142825] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.181829] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130010, 'name': PowerOffVM_Task, 'duration_secs': 0.768795} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.182569] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.182569] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 905.182569] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f5481dd-73cd-405a-8f8d-965743e7fad1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.223174] env[61991]: DEBUG nova.network.neutron [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Successfully created port: 0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.270678] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 905.271015] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 905.271329] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleting the datastore file [datastore1] 3bcde671-5702-4b8a-8881-88eb7dfd0556 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.271629] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08830bb7-6c00-4fde-9133-70da958df1b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.277894] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 905.277894] env[61991]: value = "task-1130013" [ 905.277894] env[61991]: _type = "Task" [ 905.277894] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.286327] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.322086] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.481211] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.481602] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ecad320-0b0e-4a2d-ab6f-0253c82f82f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.494030] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 905.494030] env[61991]: value = "task-1130014" [ 905.494030] env[61991]: _type = "Task" [ 905.494030] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.503667] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.560940] env[61991]: DEBUG nova.network.neutron [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 905.651130] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130011, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.661499] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a2dc60-2ba9-4ff0-b4cb-a6bfdaafe742 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.669791] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0075c0ac-9fff-47b4-8fe0-f14b24f01921 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.705901] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64abee3-7bcb-49f6-bcc0-31ab54820823 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.716380] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db80bc4b-a2d2-4939-b419-5fd7d6a270f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.730697] env[61991]: DEBUG nova.compute.provider_tree [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.788313] env[61991]: DEBUG oslo_vmware.api [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.341031} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.788571] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.788799] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.788993] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.789198] env[61991]: INFO nova.compute.manager [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 1.64 seconds to destroy the instance on the hypervisor. [ 905.789450] env[61991]: DEBUG oslo.service.loopingcall [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.789700] env[61991]: DEBUG nova.compute.manager [-] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.789809] env[61991]: DEBUG nova.network.neutron [-] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 905.822173] env[61991]: DEBUG nova.network.neutron [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Updating instance_info_cache with network_info: [{"id": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "address": "fa:16:3e:96:ee:74", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9cd766-2c", "ovs_interfaceid": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.005018] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130014, 'name': PowerOffVM_Task, 'duration_secs': 0.220211} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.005404] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.005722] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.006543] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214b57d8-85c3-4ee5-bd1e-645be16671b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.013643] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.013643] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3e06c51-1349-4933-94a5-c2ca1ba10edd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.081828] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.082130] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.082344] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore2] 0dcf415e-b554-4c82-b127-9f0e22176031 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.082637] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c524a618-9c3e-4ad6-8f71-a9395e8709ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.093157] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 906.093157] env[61991]: value = "task-1130016" [ 906.093157] env[61991]: _type = "Task" [ 906.093157] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.099527] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.144665] env[61991]: DEBUG oslo_vmware.api [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130011, 'name': PowerOnVM_Task, 'duration_secs': 0.736329} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.144970] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.145606] env[61991]: INFO nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Took 10.09 seconds to spawn the instance on the hypervisor. [ 906.145606] env[61991]: DEBUG nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.146281] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3dc3943-c110-4ffe-b9b8-5649f6ccfa0a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.256999] env[61991]: ERROR nova.scheduler.client.report [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [req-07917afa-9934-4500-a36d-f69c5bfe5bfe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-07917afa-9934-4500-a36d-f69c5bfe5bfe"}]} [ 906.274880] env[61991]: DEBUG nova.scheduler.client.report [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 906.301358] env[61991]: DEBUG nova.scheduler.client.report [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 906.301618] env[61991]: DEBUG nova.compute.provider_tree [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 906.325151] env[61991]: DEBUG nova.scheduler.client.report [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 906.325151] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.326053] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Instance network_info: |[{"id": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "address": "fa:16:3e:96:ee:74", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9cd766-2c", "ovs_interfaceid": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 906.326914] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:ee:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd2f5e5e2-e460-49ce-aa24-232e4a8007af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b9cd766-2c3d-4f0a-beb3-30531d95c2ff', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.338498] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Creating folder: Project (4c696231bd014ee5a67e4c3719f94ab4). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 906.340868] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.343728] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d53fb2b1-3533-4b38-a211-634c93649a4e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.350716] env[61991]: DEBUG nova.scheduler.client.report [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 906.358018] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Created folder: Project (4c696231bd014ee5a67e4c3719f94ab4) in parent group-v246753. [ 906.358018] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Creating folder: Instances. Parent ref: group-v246919. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 906.358018] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc4bf905-3d78-4e4b-9a30-969fea747ef0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.372659] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Created folder: Instances in parent group-v246919. [ 906.373183] env[61991]: DEBUG oslo.service.loopingcall [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.373490] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.373867] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb5addcd-f5d0-4d55-b04b-5eb6b9f21e60 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.393160] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.393553] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.393795] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.394160] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.394386] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.394614] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.394957] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.395195] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.395459] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.395754] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.396122] env[61991]: DEBUG nova.virt.hardware [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.399868] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7a38fa-f527-495e-9f17-3778239e53c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.412589] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa870a02-bc51-434e-a206-743d31814666 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.417757] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.417757] env[61991]: value = "task-1130019" [ 906.417757] env[61991]: _type = "Task" [ 906.417757] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.420557] env[61991]: DEBUG nova.compute.manager [req-96b273fa-ea12-4b02-bca9-590dc196f826 req-797dae19-3fc5-4632-b046-8e81cc2c7dc2 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Received event network-vif-deleted-dff90314-9bf7-4448-980c-047fc0a617ec {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.420972] env[61991]: INFO nova.compute.manager [req-96b273fa-ea12-4b02-bca9-590dc196f826 req-797dae19-3fc5-4632-b046-8e81cc2c7dc2 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Neutron deleted interface dff90314-9bf7-4448-980c-047fc0a617ec; detaching it from the instance and deleting it from the info cache [ 906.421236] env[61991]: DEBUG nova.network.neutron [req-96b273fa-ea12-4b02-bca9-590dc196f826 req-797dae19-3fc5-4632-b046-8e81cc2c7dc2 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.453075] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130019, 'name': CreateVM_Task} progress is 15%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.455701] env[61991]: DEBUG nova.compute.manager [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Received event network-changed-9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.456171] env[61991]: DEBUG nova.compute.manager [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Refreshing instance network info cache due to event network-changed-9b9cd766-2c3d-4f0a-beb3-30531d95c2ff. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.456523] env[61991]: DEBUG oslo_concurrency.lockutils [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] Acquiring lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.456782] env[61991]: DEBUG oslo_concurrency.lockutils [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] Acquired lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.457051] env[61991]: DEBUG nova.network.neutron [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Refreshing network info cache for port 9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.604400] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281085} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.604672] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.604862] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 906.605262] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.674935] env[61991]: INFO nova.compute.manager [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Took 33.83 seconds to build instance. [ 906.712177] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6e2227-3a00-4cae-8b79-d6582d6489f1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.720602] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e516392-9d28-45a9-8c06-eb9f57eb31ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.757098] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43234e1e-3833-451b-8ff8-a16fe87d868e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.765524] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef463924-db8e-4949-a329-9e431a9f2a80 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.782392] env[61991]: DEBUG nova.compute.provider_tree [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 906.910022] env[61991]: DEBUG nova.network.neutron [-] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.931691] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130019, 'name': CreateVM_Task, 'duration_secs': 0.420614} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.932230] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.932477] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21727bbc-66ab-4b11-bc15-f7552f2b508e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.935244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.935420] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.935772] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.936337] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd234233-574f-45de-a26b-04ffdc49dbbf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.943975] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 906.943975] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c76fd7-c882-bbb3-cb3f-3572020084d8" [ 906.943975] env[61991]: _type = "Task" [ 906.943975] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.951446] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff0c1b3-16a3-42ad-8530-02cfc31accab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.971983] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c76fd7-c882-bbb3-cb3f-3572020084d8, 'name': SearchDatastore_Task, 'duration_secs': 0.020255} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.972338] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.972573] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.972814] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.972964] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.973163] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.973434] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-810d82c4-5908-4263-862f-24b47c923dd1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.992147] env[61991]: DEBUG nova.compute.manager [req-96b273fa-ea12-4b02-bca9-590dc196f826 req-797dae19-3fc5-4632-b046-8e81cc2c7dc2 service nova] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Detach interface failed, port_id=dff90314-9bf7-4448-980c-047fc0a617ec, reason: Instance 3bcde671-5702-4b8a-8881-88eb7dfd0556 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 906.999926] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.000224] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.001084] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eec1853-dff0-4d3e-8cf9-a08505397a39 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.009313] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 907.009313] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5204b561-8c29-9312-9c41-9c39253aef23" [ 907.009313] env[61991]: _type = "Task" [ 907.009313] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.019800] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5204b561-8c29-9312-9c41-9c39253aef23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.178085] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a25df6d4-6ef0-4bfc-b3af-e7169f0ca2b6 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.348s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.247166] env[61991]: DEBUG nova.network.neutron [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Successfully updated port: 0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.320573] env[61991]: DEBUG nova.scheduler.client.report [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 907.320856] env[61991]: DEBUG nova.compute.provider_tree [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 119 to 120 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 907.321055] env[61991]: DEBUG nova.compute.provider_tree [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.413396] env[61991]: INFO nova.compute.manager [-] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Took 1.62 seconds to deallocate network for instance. [ 907.519766] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5204b561-8c29-9312-9c41-9c39253aef23, 'name': SearchDatastore_Task, 'duration_secs': 0.019194} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.520623] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50ef4767-1040-4b86-908d-fb37ab5d388f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.526294] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 907.526294] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d405b5-7e6e-9e0b-c079-8cc861e47927" [ 907.526294] env[61991]: _type = "Task" [ 907.526294] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.535875] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d405b5-7e6e-9e0b-c079-8cc861e47927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.649059] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.649286] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.649448] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.649615] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.649764] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.649909] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.650137] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.650289] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.650454] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.650621] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.650796] env[61991]: DEBUG nova.virt.hardware [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.651708] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3682a0b2-0346-4890-b32f-e570d641c7ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.661049] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be6f2eb-ebba-45a7-b818-5624409f2207 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.679105] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:1e:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b76d924-ef23-455b-afc4-7ea48006e69e', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.687894] env[61991]: DEBUG oslo.service.loopingcall [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.687894] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.687894] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54422163-88dc-4891-ac91-125ca007be63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.710376] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.710376] env[61991]: value = "task-1130020" [ 907.710376] env[61991]: _type = "Task" [ 907.710376] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.721391] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130020, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.735977] env[61991]: DEBUG nova.network.neutron [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Updated VIF entry in instance network info cache for port 9b9cd766-2c3d-4f0a-beb3-30531d95c2ff. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 907.736499] env[61991]: DEBUG nova.network.neutron [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Updating instance_info_cache with network_info: [{"id": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "address": "fa:16:3e:96:ee:74", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9cd766-2c", "ovs_interfaceid": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.748920] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "refresh_cache-a21cc5c6-abd8-45c6-b24b-33d3374df65c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.748920] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "refresh_cache-a21cc5c6-abd8-45c6-b24b-33d3374df65c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.750546] env[61991]: DEBUG nova.network.neutron [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.831021] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.524s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.831021] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.832879] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.135s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.834745] env[61991]: INFO nova.compute.claims [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.930044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.040508] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d405b5-7e6e-9e0b-c079-8cc861e47927, 'name': SearchDatastore_Task, 'duration_secs': 0.026813} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.040808] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.043018] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 69a272a7-8da6-4fb2-95be-85cfb7d70d44/69a272a7-8da6-4fb2-95be-85cfb7d70d44.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.043018] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59b14c38-8548-4840-97d2-e4c8d6c1dcdd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.051445] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 908.051445] env[61991]: value = "task-1130021" [ 908.051445] env[61991]: _type = "Task" [ 908.051445] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.064628] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130021, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.221427] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130020, 'name': CreateVM_Task, 'duration_secs': 0.480016} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.221940] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.222819] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.223242] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.223735] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.224816] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4fb1c96-c1bd-44d7-8d30-0874c53f8fea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.230178] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 908.230178] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520721cb-ca47-37fc-4324-077ea6572f56" [ 908.230178] env[61991]: _type = "Task" [ 908.230178] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.238749] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520721cb-ca47-37fc-4324-077ea6572f56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.243697] env[61991]: DEBUG oslo_concurrency.lockutils [req-18c24960-f62b-457f-af57-beb86c7f4e49 req-6915b6cb-035c-4397-98c7-0600a9052813 service nova] Releasing lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.289019] env[61991]: DEBUG nova.network.neutron [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.340412] env[61991]: DEBUG nova.compute.utils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.347983] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.347983] env[61991]: DEBUG nova.network.neutron [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 908.431973] env[61991]: DEBUG nova.policy [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be32a956d6f6471faabf7a1861800290', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c696231bd014ee5a67e4c3719f94ab4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.486870] env[61991]: DEBUG nova.network.neutron [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Updating instance_info_cache with network_info: [{"id": "0743e42e-3fb0-4809-9305-87b229303bcf", "address": "fa:16:3e:5a:8e:9c", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0743e42e-3f", "ovs_interfaceid": "0743e42e-3fb0-4809-9305-87b229303bcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.496644] env[61991]: DEBUG nova.compute.manager [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.496644] env[61991]: DEBUG nova.compute.manager [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing instance network info cache due to event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.496644] env[61991]: DEBUG oslo_concurrency.lockutils [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.496644] env[61991]: DEBUG oslo_concurrency.lockutils [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.496644] env[61991]: DEBUG nova.network.neutron [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 908.522917] env[61991]: DEBUG nova.compute.manager [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Received event network-vif-plugged-0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.523115] env[61991]: DEBUG oslo_concurrency.lockutils [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] Acquiring lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.523369] env[61991]: DEBUG oslo_concurrency.lockutils [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.523643] env[61991]: DEBUG oslo_concurrency.lockutils [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.523884] env[61991]: DEBUG nova.compute.manager [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] No waiting events found dispatching network-vif-plugged-0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.524780] env[61991]: WARNING nova.compute.manager [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Received unexpected event network-vif-plugged-0743e42e-3fb0-4809-9305-87b229303bcf for instance with vm_state building and task_state spawning. [ 908.525015] env[61991]: DEBUG nova.compute.manager [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Received event network-changed-0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.525367] env[61991]: DEBUG nova.compute.manager [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Refreshing instance network info cache due to event network-changed-0743e42e-3fb0-4809-9305-87b229303bcf. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.525711] env[61991]: DEBUG oslo_concurrency.lockutils [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] Acquiring lock "refresh_cache-a21cc5c6-abd8-45c6-b24b-33d3374df65c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.564344] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130021, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.740700] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520721cb-ca47-37fc-4324-077ea6572f56, 'name': SearchDatastore_Task, 'duration_secs': 0.013372} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.741356] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.741780] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 908.742344] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.742747] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.742963] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.743291] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f6ddbe2-04de-4afd-9f86-4ca66f991653 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.751606] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.751811] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.752572] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9365caf7-2135-4cf9-aedf-95236fc1d5eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.757794] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 908.757794] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cadfc7-095a-b359-ee5d-2e146a149e72" [ 908.757794] env[61991]: _type = "Task" [ 908.757794] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.765924] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cadfc7-095a-b359-ee5d-2e146a149e72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.848032] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.915124] env[61991]: DEBUG nova.network.neutron [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Successfully created port: 2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.982393] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e58007-796b-142f-7380-afd983384c6b/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 908.983395] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d0c81f-9876-4dfc-b385-a81709d6b4d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.990029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "refresh_cache-a21cc5c6-abd8-45c6-b24b-33d3374df65c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.990336] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Instance network_info: |[{"id": "0743e42e-3fb0-4809-9305-87b229303bcf", "address": "fa:16:3e:5a:8e:9c", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0743e42e-3f", "ovs_interfaceid": "0743e42e-3fb0-4809-9305-87b229303bcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.991182] env[61991]: DEBUG oslo_concurrency.lockutils [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] Acquired lock "refresh_cache-a21cc5c6-abd8-45c6-b24b-33d3374df65c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.991347] env[61991]: DEBUG nova.network.neutron [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Refreshing network info cache for port 0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 908.993951] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:8e:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd2f5e5e2-e460-49ce-aa24-232e4a8007af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0743e42e-3fb0-4809-9305-87b229303bcf', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.002119] env[61991]: DEBUG oslo.service.loopingcall [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.009170] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 909.009464] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e58007-796b-142f-7380-afd983384c6b/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 909.009624] env[61991]: ERROR oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e58007-796b-142f-7380-afd983384c6b/disk-0.vmdk due to incomplete transfer. [ 909.010370] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b282858-43d0-43fc-ab51-6ca9451e9c6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.034441] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4ef82694-0017-4d27-b093-9e24cba369b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.046997] env[61991]: DEBUG oslo_vmware.rw_handles [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e58007-796b-142f-7380-afd983384c6b/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 909.047485] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Uploaded image bb20df08-8e30-49f4-988e-4841fc317709 to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 909.049801] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 909.051478] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b13fbcbe-2790-48f2-b580-aeb3c3fdabdb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.053052] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.053052] env[61991]: value = "task-1130022" [ 909.053052] env[61991]: _type = "Task" [ 909.053052] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.065334] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 909.065334] env[61991]: value = "task-1130023" [ 909.065334] env[61991]: _type = "Task" [ 909.065334] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.076193] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130022, 'name': CreateVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.076596] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130021, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673231} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.080046] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 69a272a7-8da6-4fb2-95be-85cfb7d70d44/69a272a7-8da6-4fb2-95be-85cfb7d70d44.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.080170] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.080692] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01c73e54-f8a3-479e-aefb-192984aec7bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.093345] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130023, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.098090] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 909.098090] env[61991]: value = "task-1130024" [ 909.098090] env[61991]: _type = "Task" [ 909.098090] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.111251] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130024, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.223622] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48247b2-cb8e-4266-b408-0874e55532ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.235871] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8e5af5-5b5b-4e2e-affc-2e7793daed09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.278019] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c0b3c1-9789-4a66-8b2f-cecb87ca6060 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.290202] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60455e7-f800-46de-8173-c7fd80495aa9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.294517] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52cadfc7-095a-b359-ee5d-2e146a149e72, 'name': SearchDatastore_Task, 'duration_secs': 0.013802} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.295880] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-294b3d8b-148c-48c4-913c-d7c5524712e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.310423] env[61991]: DEBUG nova.compute.provider_tree [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 909.315193] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 909.315193] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5256ba83-6d42-748e-2e29-adf7b0a0f2ec" [ 909.315193] env[61991]: _type = "Task" [ 909.315193] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.324714] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5256ba83-6d42-748e-2e29-adf7b0a0f2ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.476095] env[61991]: DEBUG nova.network.neutron [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updated VIF entry in instance network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 909.476507] env[61991]: DEBUG nova.network.neutron [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.567520] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130022, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.578575] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130023, 'name': Destroy_Task, 'duration_secs': 0.325221} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.578575] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Destroyed the VM [ 909.578575] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 909.578753] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-18841aa5-c0b9-483a-902e-c43ea44960ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.585024] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 909.585024] env[61991]: value = "task-1130025" [ 909.585024] env[61991]: _type = "Task" [ 909.585024] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.593023] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130025, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.607724] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130024, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068906} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.607954] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.608846] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca44acd7-9925-499d-a301-fea89382198f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.631396] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 69a272a7-8da6-4fb2-95be-85cfb7d70d44/69a272a7-8da6-4fb2-95be-85cfb7d70d44.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.632158] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5123e97-8258-4f89-8dff-50c547666354 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.653633] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 909.653633] env[61991]: value = "task-1130026" [ 909.653633] env[61991]: _type = "Task" [ 909.653633] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.666455] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130026, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.826519] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5256ba83-6d42-748e-2e29-adf7b0a0f2ec, 'name': SearchDatastore_Task, 'duration_secs': 0.012315} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.826519] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.826519] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.827758] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c644612-b65b-4acc-8c89-e7b7c043934a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.833206] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 909.833206] env[61991]: value = "task-1130027" [ 909.833206] env[61991]: _type = "Task" [ 909.833206] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.840895] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130027, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.855397] env[61991]: DEBUG nova.scheduler.client.report [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 909.855727] env[61991]: DEBUG nova.compute.provider_tree [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 120 to 121 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 909.855929] env[61991]: DEBUG nova.compute.provider_tree [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 909.859935] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.893666] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.893939] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.894414] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.894414] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.894576] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.894825] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.895073] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.895760] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.895760] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.895760] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.895760] env[61991]: DEBUG nova.virt.hardware [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.897048] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea9de39-b236-42d1-93dc-72d75a8d9ed8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.905987] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed23eed-7845-409f-b8f9-3ae1baefd26f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.978595] env[61991]: DEBUG oslo_concurrency.lockutils [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.979242] env[61991]: DEBUG nova.compute.manager [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.979242] env[61991]: DEBUG nova.compute.manager [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing instance network info cache due to event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.979383] env[61991]: DEBUG oslo_concurrency.lockutils [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.979530] env[61991]: DEBUG oslo_concurrency.lockutils [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.979754] env[61991]: DEBUG nova.network.neutron [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.065438] env[61991]: DEBUG nova.network.neutron [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Updated VIF entry in instance network info cache for port 0743e42e-3fb0-4809-9305-87b229303bcf. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 910.065438] env[61991]: DEBUG nova.network.neutron [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Updating instance_info_cache with network_info: [{"id": "0743e42e-3fb0-4809-9305-87b229303bcf", "address": "fa:16:3e:5a:8e:9c", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0743e42e-3f", "ovs_interfaceid": "0743e42e-3fb0-4809-9305-87b229303bcf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.072283] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130022, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.098861] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130025, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.166921] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130026, 'name': ReconfigVM_Task, 'duration_secs': 0.271305} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.167243] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 69a272a7-8da6-4fb2-95be-85cfb7d70d44/69a272a7-8da6-4fb2-95be-85cfb7d70d44.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.167917] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b89fca33-9b2f-45c7-84c7-90c0050e95c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.176148] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 910.176148] env[61991]: value = "task-1130028" [ 910.176148] env[61991]: _type = "Task" [ 910.176148] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.185830] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130028, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.343348] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130027, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.367247] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.367789] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.371129] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.954s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.372827] env[61991]: INFO nova.compute.claims [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.569212] env[61991]: DEBUG oslo_concurrency.lockutils [req-986f1cb8-9ecc-478e-9ce2-96f64c0c12d6 req-ff1916d0-8c20-4966-b489-d5efbf2dacb6 service nova] Releasing lock "refresh_cache-a21cc5c6-abd8-45c6-b24b-33d3374df65c" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.569594] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130022, 'name': CreateVM_Task, 'duration_secs': 1.441334} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.569750] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.570425] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.570595] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.570916] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.571530] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c3b82e9-353b-4745-973e-62541662a939 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.576456] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 910.576456] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5245ec08-6e19-de28-27b3-82db0089e592" [ 910.576456] env[61991]: _type = "Task" [ 910.576456] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.585708] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5245ec08-6e19-de28-27b3-82db0089e592, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.593749] env[61991]: DEBUG oslo_vmware.api [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130025, 'name': RemoveSnapshot_Task, 'duration_secs': 0.519638} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.594085] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 910.594368] env[61991]: INFO nova.compute.manager [None req-66f97d47-e4ad-4f48-b4e9-c84dba46c180 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Took 14.47 seconds to snapshot the instance on the hypervisor. [ 910.604946] env[61991]: DEBUG nova.network.neutron [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Successfully updated port: 2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.652365] env[61991]: DEBUG nova.compute.manager [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.652725] env[61991]: DEBUG nova.compute.manager [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing instance network info cache due to event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 910.652808] env[61991]: DEBUG oslo_concurrency.lockutils [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.671989] env[61991]: DEBUG nova.compute.manager [req-553a371f-7be1-45e0-ae07-2a637b573444 req-0b6b888f-b9fa-42ab-95b3-b4bf64ecf05b service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Received event network-vif-plugged-2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.672236] env[61991]: DEBUG oslo_concurrency.lockutils [req-553a371f-7be1-45e0-ae07-2a637b573444 req-0b6b888f-b9fa-42ab-95b3-b4bf64ecf05b service nova] Acquiring lock "077c5def-84cc-48f9-be90-410100666773-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.672807] env[61991]: DEBUG oslo_concurrency.lockutils [req-553a371f-7be1-45e0-ae07-2a637b573444 req-0b6b888f-b9fa-42ab-95b3-b4bf64ecf05b service nova] Lock "077c5def-84cc-48f9-be90-410100666773-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.672807] env[61991]: DEBUG oslo_concurrency.lockutils [req-553a371f-7be1-45e0-ae07-2a637b573444 req-0b6b888f-b9fa-42ab-95b3-b4bf64ecf05b service nova] Lock "077c5def-84cc-48f9-be90-410100666773-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.672807] env[61991]: DEBUG nova.compute.manager [req-553a371f-7be1-45e0-ae07-2a637b573444 req-0b6b888f-b9fa-42ab-95b3-b4bf64ecf05b service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] No waiting events found dispatching network-vif-plugged-2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.672996] env[61991]: WARNING nova.compute.manager [req-553a371f-7be1-45e0-ae07-2a637b573444 req-0b6b888f-b9fa-42ab-95b3-b4bf64ecf05b service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Received unexpected event network-vif-plugged-2ca5206e-8353-4094-b7cd-257f04c3fdcb for instance with vm_state building and task_state spawning. [ 910.686829] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130028, 'name': Rename_Task, 'duration_secs': 0.150651} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.687634] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.688097] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-390562fd-9e11-4ff5-8c02-5f7353c59fe7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.698019] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 910.698019] env[61991]: value = "task-1130029" [ 910.698019] env[61991]: _type = "Task" [ 910.698019] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.703810] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.843858] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130027, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.874341] env[61991]: DEBUG nova.compute.utils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.875481] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.876354] env[61991]: DEBUG nova.network.neutron [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 910.938782] env[61991]: DEBUG nova.policy [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8c995dbd884e95bd04e034e2999a56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e4cb3d1518d498a8cdc2aee3acb90cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 910.991028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-9d1ac179-3d7e-4f0b-a80d-033eba447cab-b603ad92-a23f-4da6-b5dc-60d547ba32ae" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.991028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-9d1ac179-3d7e-4f0b-a80d-033eba447cab-b603ad92-a23f-4da6-b5dc-60d547ba32ae" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.991028] env[61991]: DEBUG nova.objects.instance [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'flavor' on Instance uuid 9d1ac179-3d7e-4f0b-a80d-033eba447cab {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.992364] env[61991]: DEBUG nova.network.neutron [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updated VIF entry in instance network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 910.992964] env[61991]: DEBUG nova.network.neutron [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.089023] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5245ec08-6e19-de28-27b3-82db0089e592, 'name': SearchDatastore_Task, 'duration_secs': 0.010836} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.089023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.089023] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.089023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.089023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.089023] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.089023] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0390f02-e0b4-4e10-8493-e914e9d29bc1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.095705] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.096157] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.099681] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e38bfa8c-cfa4-4d2b-93a3-f927fae8f243 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.108963] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 911.108963] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a4f595-a76f-399b-e460-a985808e3df1" [ 911.108963] env[61991]: _type = "Task" [ 911.108963] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.113861] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "refresh_cache-077c5def-84cc-48f9-be90-410100666773" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.113861] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "refresh_cache-077c5def-84cc-48f9-be90-410100666773" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.113861] env[61991]: DEBUG nova.network.neutron [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.119856] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a4f595-a76f-399b-e460-a985808e3df1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.205513] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130029, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.254541] env[61991]: DEBUG nova.network.neutron [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Successfully created port: 02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.347750] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130027, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.2787} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.347750] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.347750] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.347750] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-340d21a3-4461-4dcf-aa69-b04764275161 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.354641] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 911.354641] env[61991]: value = "task-1130030" [ 911.354641] env[61991]: _type = "Task" [ 911.354641] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.366473] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130030, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.380135] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.502482] env[61991]: DEBUG oslo_concurrency.lockutils [req-70693350-dfb3-46cd-99c2-f6bf40c361eb req-7c66d293-c263-434b-b817-b7293a8c53f5 service nova] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.503243] env[61991]: DEBUG oslo_concurrency.lockutils [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.503523] env[61991]: DEBUG nova.network.neutron [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.622245] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a4f595-a76f-399b-e460-a985808e3df1, 'name': SearchDatastore_Task, 'duration_secs': 0.014321} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.622245] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd59e6a4-8a59-497e-af33-e0ba9712df7f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.632875] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 911.632875] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520c64c8-615f-9aad-0ec0-b57e96430598" [ 911.632875] env[61991]: _type = "Task" [ 911.632875] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.637767] env[61991]: DEBUG nova.objects.instance [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'pci_requests' on Instance uuid 9d1ac179-3d7e-4f0b-a80d-033eba447cab {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.645188] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520c64c8-615f-9aad-0ec0-b57e96430598, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.662595] env[61991]: DEBUG nova.network.neutron [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 911.672422] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d4e3c6-54ed-4d43-8de2-a6af433e8a52 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.681340] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42db1283-f748-41b0-94e7-ed92c746d35d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.716311] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628a1ba1-5241-4643-8222-9adfeb56e792 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.724986] env[61991]: DEBUG oslo_vmware.api [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130029, 'name': PowerOnVM_Task, 'duration_secs': 0.937474} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.726142] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.726390] env[61991]: INFO nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Took 7.79 seconds to spawn the instance on the hypervisor. [ 911.726535] env[61991]: DEBUG nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.727379] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa487a6c-0a0e-411c-b906-5b7499c40b7b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.730674] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24763ce4-c3be-40ff-9c1f-a405b2c703fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.749095] env[61991]: DEBUG nova.compute.provider_tree [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.836863] env[61991]: DEBUG nova.network.neutron [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Updating instance_info_cache with network_info: [{"id": "2ca5206e-8353-4094-b7cd-257f04c3fdcb", "address": "fa:16:3e:3c:b1:d5", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ca5206e-83", "ovs_interfaceid": "2ca5206e-8353-4094-b7cd-257f04c3fdcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.867018] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130030, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103359} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.867018] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.867018] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f6ea9d-c048-4db6-8345-8d0416777ca7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.888688] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.892667] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc21448e-7706-47f5-acce-67c7de200751 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.911729] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 911.911729] env[61991]: value = "task-1130031" [ 911.911729] env[61991]: _type = "Task" [ 911.911729] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.919664] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.139995] env[61991]: DEBUG nova.objects.base [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Object Instance<9d1ac179-3d7e-4f0b-a80d-033eba447cab> lazy-loaded attributes: flavor,pci_requests {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 912.140281] env[61991]: DEBUG nova.network.neutron [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.146214] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520c64c8-615f-9aad-0ec0-b57e96430598, 'name': SearchDatastore_Task, 'duration_secs': 0.018285} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.146737] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.147045] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] a21cc5c6-abd8-45c6-b24b-33d3374df65c/a21cc5c6-abd8-45c6-b24b-33d3374df65c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.147504] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d537beb9-cd03-4400-a4e2-ec06d8ee318d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.153630] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 912.153630] env[61991]: value = "task-1130032" [ 912.153630] env[61991]: _type = "Task" [ 912.153630] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.161389] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.211640] env[61991]: DEBUG nova.network.neutron [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updated VIF entry in instance network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.212282] env[61991]: DEBUG nova.network.neutron [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.223256] env[61991]: DEBUG nova.policy [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.260149] env[61991]: INFO nova.compute.manager [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Took 23.57 seconds to build instance. [ 912.287053] env[61991]: DEBUG nova.scheduler.client.report [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 121 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 912.287053] env[61991]: DEBUG nova.compute.provider_tree [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 121 to 122 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 912.287053] env[61991]: DEBUG nova.compute.provider_tree [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.339793] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "refresh_cache-077c5def-84cc-48f9-be90-410100666773" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.339793] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Instance network_info: |[{"id": "2ca5206e-8353-4094-b7cd-257f04c3fdcb", "address": "fa:16:3e:3c:b1:d5", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ca5206e-83", "ovs_interfaceid": "2ca5206e-8353-4094-b7cd-257f04c3fdcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.340856] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:b1:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd2f5e5e2-e460-49ce-aa24-232e4a8007af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ca5206e-8353-4094-b7cd-257f04c3fdcb', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.347917] env[61991]: DEBUG oslo.service.loopingcall [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.348163] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 077c5def-84cc-48f9-be90-410100666773] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.348382] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5987451f-57e5-4c49-af7b-afb70248b239 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.369250] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.369250] env[61991]: value = "task-1130033" [ 912.369250] env[61991]: _type = "Task" [ 912.369250] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.376756] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130033, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.408379] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.423494] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130031, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.431376] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.431616] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.431775] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.431965] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.432117] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.432286] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.432462] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.432620] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.432789] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.432954] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.433148] env[61991]: DEBUG nova.virt.hardware [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.433943] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a794d3b-0ff6-473e-9e1d-1b1f14314632 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.441029] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bfc84e-560e-4a43-a4fd-45b636ad9342 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.669882] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130032, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.716295] env[61991]: DEBUG oslo_concurrency.lockutils [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.716612] env[61991]: DEBUG nova.compute.manager [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.716900] env[61991]: DEBUG nova.compute.manager [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing instance network info cache due to event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.717062] env[61991]: DEBUG oslo_concurrency.lockutils [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.717535] env[61991]: DEBUG oslo_concurrency.lockutils [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.717535] env[61991]: DEBUG nova.network.neutron [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.762779] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9ef78d9c-4332-4985-8913-235a9c68e701 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.078s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.789921] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.790623] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.794476] env[61991]: DEBUG nova.network.neutron [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Successfully updated port: 02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.798042] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.938s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.798042] env[61991]: DEBUG nova.objects.instance [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lazy-loading 'resources' on Instance uuid 39004ce4-c96b-4911-9b64-3c508ddcecaf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.881836] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130033, 'name': CreateVM_Task, 'duration_secs': 0.358628} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.882087] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 077c5def-84cc-48f9-be90-410100666773] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.882757] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.882968] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.883268] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.883527] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-501908e8-f0f8-4e41-876c-1aac7f2b3427 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.888681] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 912.888681] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520b08b7-17c4-9e45-c852-cc7948cf33cc" [ 912.888681] env[61991]: _type = "Task" [ 912.888681] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.897356] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520b08b7-17c4-9e45-c852-cc7948cf33cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.922242] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130031, 'name': ReconfigVM_Task, 'duration_secs': 0.96879} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.922596] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 0dcf415e-b554-4c82-b127-9f0e22176031/0dcf415e-b554-4c82-b127-9f0e22176031.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.923302] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a46b8080-4a47-4522-a5f3-b5f84b8a68a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.929675] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 912.929675] env[61991]: value = "task-1130034" [ 912.929675] env[61991]: _type = "Task" [ 912.929675] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.938105] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130034, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.963033] env[61991]: DEBUG nova.compute.manager [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Received event network-changed-2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.963287] env[61991]: DEBUG nova.compute.manager [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Refreshing instance network info cache due to event network-changed-2ca5206e-8353-4094-b7cd-257f04c3fdcb. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.963543] env[61991]: DEBUG oslo_concurrency.lockutils [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] Acquiring lock "refresh_cache-077c5def-84cc-48f9-be90-410100666773" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.963711] env[61991]: DEBUG oslo_concurrency.lockutils [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] Acquired lock "refresh_cache-077c5def-84cc-48f9-be90-410100666773" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.963912] env[61991]: DEBUG nova.network.neutron [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Refreshing network info cache for port 2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.085255] env[61991]: DEBUG nova.compute.manager [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.086231] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7888fef-01d8-4aea-8f2b-7c9a143a4bf7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.163217] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130032, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721672} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.163446] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] a21cc5c6-abd8-45c6-b24b-33d3374df65c/a21cc5c6-abd8-45c6-b24b-33d3374df65c.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.163662] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.163918] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d722f802-f9e7-4a78-81d4-3dc81b5c7640 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.170065] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 913.170065] env[61991]: value = "task-1130035" [ 913.170065] env[61991]: _type = "Task" [ 913.170065] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.179111] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130035, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.300435] env[61991]: DEBUG nova.compute.utils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.302038] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.302038] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.302038] env[61991]: DEBUG nova.network.neutron [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.307345] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.307574] env[61991]: DEBUG nova.network.neutron [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 913.404246] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520b08b7-17c4-9e45-c852-cc7948cf33cc, 'name': SearchDatastore_Task, 'duration_secs': 0.008955} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.404702] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.404800] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.405056] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.405213] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.405399] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.408130] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a93d3bf-59b1-4a07-9ad7-7087ddcc4b90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.416806] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.416806] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.417488] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de777e5d-b213-4ea0-9d55-c2c375cf7b7b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.423970] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 913.423970] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c0a4f2-06eb-3c40-8585-b24c03da30bb" [ 913.423970] env[61991]: _type = "Task" [ 913.423970] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.432603] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c0a4f2-06eb-3c40-8585-b24c03da30bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.442524] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130034, 'name': Rename_Task, 'duration_secs': 0.129157} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.444751] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.445206] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0a405e4-1220-4733-a1a3-e10f9acbc03b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.454753] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 913.454753] env[61991]: value = "task-1130036" [ 913.454753] env[61991]: _type = "Task" [ 913.454753] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.468658] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.583513] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a101ad9a-113b-42a0-8358-aec48787f0f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.590700] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f43ce1e-7a89-4146-b866-5dcea355c41f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.595094] env[61991]: DEBUG nova.policy [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e647f3055e254e2d9b48f976e710348c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c89707d8b26430c830449ab9bca4a62', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.597241] env[61991]: INFO nova.compute.manager [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] instance snapshotting [ 913.601445] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862703b4-0fcf-4fa9-b014-7ad06cbcaade {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.633646] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b11338-e497-481d-966b-5ea8f948d6c0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.649822] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42482963-3943-47bd-b9ba-9409e3525c1b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.657757] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a15d6c-a16f-4d0a-b46b-8cfa946f8859 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.678201] env[61991]: DEBUG nova.compute.provider_tree [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.686801] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130035, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063245} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.687096] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.688453] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e058b90f-2ee8-4191-ac68-7d801c6a63f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.710925] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] a21cc5c6-abd8-45c6-b24b-33d3374df65c/a21cc5c6-abd8-45c6-b24b-33d3374df65c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.713864] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a07a76c-8157-4023-818d-748854695200 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.737258] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 913.737258] env[61991]: value = "task-1130037" [ 913.737258] env[61991]: _type = "Task" [ 913.737258] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.745853] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130037, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.803134] env[61991]: DEBUG nova.network.neutron [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updated VIF entry in instance network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.803529] env[61991]: DEBUG nova.network.neutron [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.804972] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.857522] env[61991]: DEBUG nova.network.neutron [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.882199] env[61991]: DEBUG nova.network.neutron [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Updated VIF entry in instance network info cache for port 2ca5206e-8353-4094-b7cd-257f04c3fdcb. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.882558] env[61991]: DEBUG nova.network.neutron [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Updating instance_info_cache with network_info: [{"id": "2ca5206e-8353-4094-b7cd-257f04c3fdcb", "address": "fa:16:3e:3c:b1:d5", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ca5206e-83", "ovs_interfaceid": "2ca5206e-8353-4094-b7cd-257f04c3fdcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.934973] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c0a4f2-06eb-3c40-8585-b24c03da30bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009549} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.939930] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b347f21-6412-45a5-979a-9a60df61579f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.947278] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 913.947278] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523bf368-4c57-8f96-2947-d3f6883794ac" [ 913.947278] env[61991]: _type = "Task" [ 913.947278] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.955215] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523bf368-4c57-8f96-2947-d3f6883794ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.970473] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130036, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.076328] env[61991]: DEBUG nova.network.neutron [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Successfully created port: 2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.076700] env[61991]: DEBUG nova.network.neutron [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updating instance_info_cache with network_info: [{"id": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "address": "fa:16:3e:2e:cc:6a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02083cbb-c0", "ovs_interfaceid": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.090900] env[61991]: DEBUG nova.network.neutron [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Successfully updated port: b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.170988] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 914.170988] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-81a1e0b9-b343-4603-99a8-3e9537234fa2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.178703] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 914.178703] env[61991]: value = "task-1130038" [ 914.178703] env[61991]: _type = "Task" [ 914.178703] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.183045] env[61991]: DEBUG nova.scheduler.client.report [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.191928] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130038, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.203637] env[61991]: DEBUG nova.compute.manager [req-c7810061-2149-405d-8886-4546e1850987 req-abec6e24-aebf-4f5e-b71a-31b99bc9e101 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-vif-plugged-b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.203819] env[61991]: DEBUG oslo_concurrency.lockutils [req-c7810061-2149-405d-8886-4546e1850987 req-abec6e24-aebf-4f5e-b71a-31b99bc9e101 service nova] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.204163] env[61991]: DEBUG oslo_concurrency.lockutils [req-c7810061-2149-405d-8886-4546e1850987 req-abec6e24-aebf-4f5e-b71a-31b99bc9e101 service nova] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.204466] env[61991]: DEBUG oslo_concurrency.lockutils [req-c7810061-2149-405d-8886-4546e1850987 req-abec6e24-aebf-4f5e-b71a-31b99bc9e101 service nova] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.204772] env[61991]: DEBUG nova.compute.manager [req-c7810061-2149-405d-8886-4546e1850987 req-abec6e24-aebf-4f5e-b71a-31b99bc9e101 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] No waiting events found dispatching network-vif-plugged-b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.205086] env[61991]: WARNING nova.compute.manager [req-c7810061-2149-405d-8886-4546e1850987 req-abec6e24-aebf-4f5e-b71a-31b99bc9e101 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received unexpected event network-vif-plugged-b603ad92-a23f-4da6-b5dc-60d547ba32ae for instance with vm_state active and task_state None. [ 914.251676] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130037, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.317225] env[61991]: DEBUG oslo_concurrency.lockutils [req-4dc02f95-ad1b-4fae-bec6-4f8ee79ff619 req-2a87b1d1-2b85-40ad-9cc3-f49f479b5bdc service nova] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.384888] env[61991]: DEBUG oslo_concurrency.lockutils [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] Releasing lock "refresh_cache-077c5def-84cc-48f9-be90-410100666773" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.385233] env[61991]: DEBUG nova.compute.manager [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Received event network-vif-plugged-02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.385443] env[61991]: DEBUG oslo_concurrency.lockutils [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.385743] env[61991]: DEBUG oslo_concurrency.lockutils [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.385944] env[61991]: DEBUG oslo_concurrency.lockutils [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.386144] env[61991]: DEBUG nova.compute.manager [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] No waiting events found dispatching network-vif-plugged-02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.386330] env[61991]: WARNING nova.compute.manager [req-b7b02be9-b5e1-42aa-b615-a228d3f6ad9d req-9925921a-277d-4c34-baff-5e2c5b43c087 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Received unexpected event network-vif-plugged-02083cbb-c0d3-414b-9882-6260c77dfd0c for instance with vm_state building and task_state spawning. [ 914.461143] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523bf368-4c57-8f96-2947-d3f6883794ac, 'name': SearchDatastore_Task, 'duration_secs': 0.04361} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.465075] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.465414] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 077c5def-84cc-48f9-be90-410100666773/077c5def-84cc-48f9-be90-410100666773.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.465771] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aaf6f9f0-79c4-4d0a-b89f-978fac4d524a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.474058] env[61991]: DEBUG oslo_vmware.api [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130036, 'name': PowerOnVM_Task, 'duration_secs': 1.016138} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.475695] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.475695] env[61991]: DEBUG nova.compute.manager [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.476075] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 914.476075] env[61991]: value = "task-1130039" [ 914.476075] env[61991]: _type = "Task" [ 914.476075] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.476829] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243b17f3-a36f-40f6-befd-29b9b0b50285 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.487552] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.580031] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.580031] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Instance network_info: |[{"id": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "address": "fa:16:3e:2e:cc:6a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02083cbb-c0", "ovs_interfaceid": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.580340] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:cc:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02083cbb-c0d3-414b-9882-6260c77dfd0c', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.592783] env[61991]: DEBUG oslo.service.loopingcall [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.593448] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.594739] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.594739] env[61991]: DEBUG nova.network.neutron [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.595890] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.596445] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56563835-148e-400d-8637-0b9986ddd8a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.621490] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.621490] env[61991]: value = "task-1130040" [ 914.621490] env[61991]: _type = "Task" [ 914.621490] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.629846] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130040, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.650420] env[61991]: WARNING nova.network.neutron [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] e6f9d471-4f42-4781-93c3-f4a9ff84317b already exists in list: networks containing: ['e6f9d471-4f42-4781-93c3-f4a9ff84317b']. ignoring it [ 914.692371] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.896s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.695898] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130038, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.700029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.645s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.702444] env[61991]: INFO nova.compute.claims [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.738739] env[61991]: INFO nova.scheduler.client.report [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocations for instance 39004ce4-c96b-4911-9b64-3c508ddcecaf [ 914.757741] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130037, 'name': ReconfigVM_Task, 'duration_secs': 0.695228} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.757741] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Reconfigured VM instance instance-00000059 to attach disk [datastore1] a21cc5c6-abd8-45c6-b24b-33d3374df65c/a21cc5c6-abd8-45c6-b24b-33d3374df65c.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.758707] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c44f42a-0eee-4acc-b81d-e95eca743da2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.768057] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 914.768057] env[61991]: value = "task-1130041" [ 914.768057] env[61991]: _type = "Task" [ 914.768057] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.778219] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130041, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.820718] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.849833] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.850377] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.850649] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.850917] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.851178] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.851391] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.851707] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.851940] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.852184] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.852572] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.852722] env[61991]: DEBUG nova.virt.hardware [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.853893] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5798bf32-a69b-43da-a119-916cd7bbf9b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.865386] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44403a6c-b8b3-47f4-94d6-b03de4b0f066 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.989079] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130039, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.001159] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.131719] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130040, 'name': CreateVM_Task, 'duration_secs': 0.405973} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.132317] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.133278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.133614] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.134047] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.134401] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52066dde-814e-400d-a91e-03b3df6c3114 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.142749] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 915.142749] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520713a4-0efc-10f9-764e-59e470319fec" [ 915.142749] env[61991]: _type = "Task" [ 915.142749] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.154469] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520713a4-0efc-10f9-764e-59e470319fec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.155816] env[61991]: DEBUG nova.network.neutron [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "address": "fa:16:3e:2d:ad:50", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb603ad92-a2", "ovs_interfaceid": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.190010] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130038, 'name': CreateSnapshot_Task, 'duration_secs': 0.882874} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.190313] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 915.191066] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62bbc474-69c6-40b2-bb4f-e692458b4fac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.254075] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ddce2bc7-a878-4597-9662-56ef8ddfdb42 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "39004ce4-c96b-4911-9b64-3c508ddcecaf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.008s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.278695] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130041, 'name': Rename_Task, 'duration_secs': 0.337555} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.278993] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.279258] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c8c8478-a3b4-47e1-aae7-268538f52977 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.285797] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 915.285797] env[61991]: value = "task-1130042" [ 915.285797] env[61991]: _type = "Task" [ 915.285797] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.293674] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.295667] env[61991]: DEBUG nova.compute.manager [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Received event network-changed-02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.295867] env[61991]: DEBUG nova.compute.manager [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Refreshing instance network info cache due to event network-changed-02083cbb-c0d3-414b-9882-6260c77dfd0c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.296100] env[61991]: DEBUG oslo_concurrency.lockutils [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] Acquiring lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.296262] env[61991]: DEBUG oslo_concurrency.lockutils [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] Acquired lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.296404] env[61991]: DEBUG nova.network.neutron [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Refreshing network info cache for port 02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 915.490164] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525925} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.490433] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 077c5def-84cc-48f9-be90-410100666773/077c5def-84cc-48f9-be90-410100666773.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.491334] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.491334] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ffdbddb7-06d3-49b2-9203-c6390c679587 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.499319] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 915.499319] env[61991]: value = "task-1130043" [ 915.499319] env[61991]: _type = "Task" [ 915.499319] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.507095] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.578545] env[61991]: DEBUG nova.network.neutron [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Successfully updated port: 2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.652346] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520713a4-0efc-10f9-764e-59e470319fec, 'name': SearchDatastore_Task, 'duration_secs': 0.014076} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.652658] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.652938] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.653201] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.653353] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.653539] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.653793] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-311d3e99-6098-438c-9352-bd0d1cec1a10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.658442] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.659066] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.659232] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.659951] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1475113-bcf1-415d-894e-11eaa3c14427 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.675779] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.676028] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.676200] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.676389] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.676539] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.676726] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.676947] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.677121] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.677290] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.677456] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.677628] env[61991]: DEBUG nova.virt.hardware [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.683732] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Reconfiguring VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 915.684829] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebe5b18a-9ff0-4df8-9de4-79a874b6494c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.696752] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.696925] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.697624] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45dc104b-6b8b-4232-a8b0-1287978f3d0c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.707897] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 915.710047] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-242b931e-938a-4011-9a19-d8140a463ac9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.712771] env[61991]: DEBUG oslo_vmware.api [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 915.712771] env[61991]: value = "task-1130044" [ 915.712771] env[61991]: _type = "Task" [ 915.712771] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.713029] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 915.713029] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c6e13d-5930-64b0-7e2d-af813b200ce9" [ 915.713029] env[61991]: _type = "Task" [ 915.713029] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.723854] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 915.723854] env[61991]: value = "task-1130045" [ 915.723854] env[61991]: _type = "Task" [ 915.723854] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.738702] env[61991]: DEBUG oslo_vmware.api [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130044, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.738702] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c6e13d-5930-64b0-7e2d-af813b200ce9, 'name': SearchDatastore_Task} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.738702] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa761e55-1fce-442f-9018-2ea4c9b1039e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.749794] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130045, 'name': CloneVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.750102] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 915.750102] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c17771-4c2e-95e4-0528-1036f1410a3b" [ 915.750102] env[61991]: _type = "Task" [ 915.750102] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.763638] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c17771-4c2e-95e4-0528-1036f1410a3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.795777] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130042, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.995638] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f46c1d5-a011-4527-ba33-ff218e536310 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.010147] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7939aa-6beb-4067-add9-d515f57fa33e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.018385] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063036} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.019011] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.019768] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206aadef-1ad7-4c09-ad2e-ec09d555db9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.057732] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a476578-a7fd-4503-b85a-4d45e36b7b40 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.057732] env[61991]: DEBUG nova.network.neutron [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updated VIF entry in instance network info cache for port 02083cbb-c0d3-414b-9882-6260c77dfd0c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 916.058192] env[61991]: DEBUG nova.network.neutron [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updating instance_info_cache with network_info: [{"id": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "address": "fa:16:3e:2e:cc:6a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02083cbb-c0", "ovs_interfaceid": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.076933] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 077c5def-84cc-48f9-be90-410100666773/077c5def-84cc-48f9-be90-410100666773.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.077845] env[61991]: DEBUG oslo_concurrency.lockutils [req-f812b4ab-65f8-48a6-ade8-381ef7e846ac req-ec8f5f4a-ae32-4ecf-83f6-bacebb7963eb service nova] Releasing lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.078569] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e5788ed-faa7-4245-a9a9-e31aff05d51a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.096403] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.096552] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.096806] env[61991]: DEBUG nova.network.neutron [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.099294] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e77988-4c97-4bff-ae3e-9a943104adc3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.105364] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 916.105364] env[61991]: value = "task-1130046" [ 916.105364] env[61991]: _type = "Task" [ 916.105364] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.116760] env[61991]: DEBUG nova.compute.provider_tree [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.124846] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130046, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.226215] env[61991]: DEBUG oslo_vmware.api [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130044, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.234887] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130045, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.260842] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c17771-4c2e-95e4-0528-1036f1410a3b, 'name': SearchDatastore_Task, 'duration_secs': 0.013078} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.261208] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.261505] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 0745dd19-6b08-447a-8333-047aeaa1fa83/0745dd19-6b08-447a-8333-047aeaa1fa83.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.261792] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-677028bb-aa08-406d-92cb-82c3ebd8dfe0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.269526] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 916.269526] env[61991]: value = "task-1130047" [ 916.269526] env[61991]: _type = "Task" [ 916.269526] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.279645] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130047, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.295798] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130042, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.434035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.434170] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.620145] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130046, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.620145] env[61991]: DEBUG nova.scheduler.client.report [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.634976] env[61991]: DEBUG nova.compute.manager [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-changed-b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.635228] env[61991]: DEBUG nova.compute.manager [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing instance network info cache due to event network-changed-b603ad92-a23f-4da6-b5dc-60d547ba32ae. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.635416] env[61991]: DEBUG oslo_concurrency.lockutils [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.635925] env[61991]: DEBUG oslo_concurrency.lockutils [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.635925] env[61991]: DEBUG nova.network.neutron [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing network info cache for port b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.644082] env[61991]: DEBUG nova.network.neutron [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.730310] env[61991]: DEBUG oslo_vmware.api [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130044, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.739282] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130045, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.782567] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130047, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.798445] env[61991]: DEBUG oslo_vmware.api [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130042, 'name': PowerOnVM_Task, 'duration_secs': 1.050593} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.798751] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.798975] env[61991]: INFO nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Took 10.46 seconds to spawn the instance on the hypervisor. [ 916.799177] env[61991]: DEBUG nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.800019] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30ccaa8-d386-40a0-94c8-48c547ecce23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.815901] env[61991]: DEBUG nova.network.neutron [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.937078] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 917.117969] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130046, 'name': ReconfigVM_Task, 'duration_secs': 0.553301} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.117969] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 077c5def-84cc-48f9-be90-410100666773/077c5def-84cc-48f9-be90-410100666773.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.118563] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6b2556f-cb5a-427f-9fde-7ca74ac82225 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.125324] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.125832] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.128553] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 917.128553] env[61991]: value = "task-1130048" [ 917.128553] env[61991]: _type = "Task" [ 917.128553] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.128845] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.199s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.129687] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.131502] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.130s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.131502] env[61991]: DEBUG nova.objects.instance [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 917.150476] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130048, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.161664] env[61991]: INFO nova.scheduler.client.report [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted allocations for instance 3bcde671-5702-4b8a-8881-88eb7dfd0556 [ 917.227653] env[61991]: DEBUG oslo_vmware.api [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130044, 'name': ReconfigVM_Task, 'duration_secs': 1.193042} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.228262] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.228506] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Reconfigured VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 917.240507] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130045, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.281716] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130047, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556148} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.284424] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 0745dd19-6b08-447a-8333-047aeaa1fa83/0745dd19-6b08-447a-8333-047aeaa1fa83.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.284693] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.284976] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a864a2c5-f933-4baf-bca6-de4a7f37c61e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.291769] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 917.291769] env[61991]: value = "task-1130049" [ 917.291769] env[61991]: _type = "Task" [ 917.291769] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.303441] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.317291] env[61991]: INFO nova.compute.manager [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Took 28.17 seconds to build instance. [ 917.318477] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.318845] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Instance network_info: |[{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.319493] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:2f:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fc2099f-0ea4-4665-948a-01d944f9275a', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.327537] env[61991]: DEBUG oslo.service.loopingcall [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.330740] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.331029] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9523ea76-36a3-409f-b9e4-651e6333e8af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.353766] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.353766] env[61991]: value = "task-1130050" [ 917.353766] env[61991]: _type = "Task" [ 917.353766] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.361100] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130050, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.417074] env[61991]: DEBUG nova.network.neutron [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updated VIF entry in instance network info cache for port b603ad92-a23f-4da6-b5dc-60d547ba32ae. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.417537] env[61991]: DEBUG nova.network.neutron [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "address": "fa:16:3e:2d:ad:50", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb603ad92-a2", "ovs_interfaceid": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.456872] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.512639] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "0dcf415e-b554-4c82-b127-9f0e22176031" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.512920] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "0dcf415e-b554-4c82-b127-9f0e22176031" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.513156] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "0dcf415e-b554-4c82-b127-9f0e22176031-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.513343] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "0dcf415e-b554-4c82-b127-9f0e22176031-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.513598] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "0dcf415e-b554-4c82-b127-9f0e22176031-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.515976] env[61991]: INFO nova.compute.manager [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Terminating instance [ 917.518056] env[61991]: DEBUG nova.compute.manager [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.518275] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.519131] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffdc3f0-35e8-41c5-8587-84277bbbf631 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.526718] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.526939] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f13b9160-0559-45a9-947d-941cb88d3293 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.533568] env[61991]: DEBUG oslo_vmware.api [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 917.533568] env[61991]: value = "task-1130051" [ 917.533568] env[61991]: _type = "Task" [ 917.533568] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.541464] env[61991]: DEBUG oslo_vmware.api [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130051, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.612322] env[61991]: DEBUG nova.compute.manager [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received event network-vif-plugged-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.612322] env[61991]: DEBUG oslo_concurrency.lockutils [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.612468] env[61991]: DEBUG oslo_concurrency.lockutils [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.612716] env[61991]: DEBUG oslo_concurrency.lockutils [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.612953] env[61991]: DEBUG nova.compute.manager [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] No waiting events found dispatching network-vif-plugged-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.613524] env[61991]: WARNING nova.compute.manager [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received unexpected event network-vif-plugged-2fc2099f-0ea4-4665-948a-01d944f9275a for instance with vm_state building and task_state spawning. [ 917.613524] env[61991]: DEBUG nova.compute.manager [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.613788] env[61991]: DEBUG nova.compute.manager [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing instance network info cache due to event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.614395] env[61991]: DEBUG oslo_concurrency.lockutils [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.614565] env[61991]: DEBUG oslo_concurrency.lockutils [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.614810] env[61991]: DEBUG nova.network.neutron [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.635919] env[61991]: DEBUG nova.compute.utils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.642710] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.642710] env[61991]: DEBUG nova.network.neutron [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 917.654935] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130048, 'name': Rename_Task, 'duration_secs': 0.15734} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.654935] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.654935] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e778c96-9ee0-43cc-8ae2-87da10e25d5b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.670049] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 917.670049] env[61991]: value = "task-1130052" [ 917.670049] env[61991]: _type = "Task" [ 917.670049] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.671042] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2846354d-6d46-406a-a048-e6c51274031e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "3bcde671-5702-4b8a-8881-88eb7dfd0556" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.523s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.683358] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130052, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.694270] env[61991]: DEBUG nova.policy [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8f218c7f19c4f98b4afe4d6fe94a5b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06f21c8f25694d4d93dfd78bf10e539c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.738650] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b937fad4-4c86-4404-b02b-de763fba12b4 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-9d1ac179-3d7e-4f0b-a80d-033eba447cab-b603ad92-a23f-4da6-b5dc-60d547ba32ae" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.749s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.746377] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130045, 'name': CloneVM_Task, 'duration_secs': 1.66998} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.746734] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Created linked-clone VM from snapshot [ 917.747787] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51c3b42-6249-499c-a6f5-8d498707a529 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.759597] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Uploading image 9a5d17c0-8aa4-41d4-8548-25843981601a {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 917.786767] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 917.786767] env[61991]: value = "vm-246927" [ 917.786767] env[61991]: _type = "VirtualMachine" [ 917.786767] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 917.787091] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-45b5603c-0489-40d8-8817-5193d72fe770 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.797970] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lease: (returnval){ [ 917.797970] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280058e-e32c-606e-10ac-3d2a1c39dd00" [ 917.797970] env[61991]: _type = "HttpNfcLease" [ 917.797970] env[61991]: } obtained for exporting VM: (result){ [ 917.797970] env[61991]: value = "vm-246927" [ 917.797970] env[61991]: _type = "VirtualMachine" [ 917.797970] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 917.798390] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the lease: (returnval){ [ 917.798390] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280058e-e32c-606e-10ac-3d2a1c39dd00" [ 917.798390] env[61991]: _type = "HttpNfcLease" [ 917.798390] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 917.805155] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066841} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.805816] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.806588] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58262e50-7cbc-4c04-9d48-641beefe4693 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.810484] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.810484] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280058e-e32c-606e-10ac-3d2a1c39dd00" [ 917.810484] env[61991]: _type = "HttpNfcLease" [ 917.810484] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 917.822567] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b87674ab-4709-444c-b976-0d0e223727f6 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.680s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.831504] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 0745dd19-6b08-447a-8333-047aeaa1fa83/0745dd19-6b08-447a-8333-047aeaa1fa83.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.832174] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbaf0e9d-2e49-4b62-af8b-1f318ef16314 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.852771] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 917.852771] env[61991]: value = "task-1130054" [ 917.852771] env[61991]: _type = "Task" [ 917.852771] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.864199] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130054, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.867452] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130050, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.920671] env[61991]: DEBUG oslo_concurrency.lockutils [req-da3aafb7-1213-48fa-94a0-64f0a29f20f0 req-356ab495-9764-4670-a2ff-d99c5034abfe service nova] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.982726] env[61991]: DEBUG nova.network.neutron [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Successfully created port: 7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.044814] env[61991]: DEBUG oslo_vmware.api [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130051, 'name': PowerOffVM_Task, 'duration_secs': 0.49295} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.045257] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.045554] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.046803] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0d6525f-e0bc-4abf-9279-0eeb92c82ae4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.070335] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "68268135-4cf4-4732-9104-1720bbb21acf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.071273] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.110766] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.110766] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.110766] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore1] 0dcf415e-b554-4c82-b127-9f0e22176031 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.110766] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f848e66c-1eee-4924-8de9-70a52c3b89d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.121244] env[61991]: DEBUG oslo_vmware.api [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 918.121244] env[61991]: value = "task-1130056" [ 918.121244] env[61991]: _type = "Task" [ 918.121244] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.136698] env[61991]: DEBUG oslo_vmware.api [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.145712] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.150960] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a925aa4d-1437-4172-93b3-577ca5cf44bc tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.152292] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.696s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.153503] env[61991]: INFO nova.compute.claims [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.182880] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130052, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.309022] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.309022] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280058e-e32c-606e-10ac-3d2a1c39dd00" [ 918.309022] env[61991]: _type = "HttpNfcLease" [ 918.309022] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 918.309022] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 918.309022] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5280058e-e32c-606e-10ac-3d2a1c39dd00" [ 918.309022] env[61991]: _type = "HttpNfcLease" [ 918.309022] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 918.310074] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1772b0a6-1a5c-442d-8131-0500874227e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.319810] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f492ab-3240-a946-852b-d64628acbe51/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 918.320267] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f492ab-3240-a946-852b-d64628acbe51/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 918.402278] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130054, 'name': ReconfigVM_Task, 'duration_secs': 0.333432} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.405558] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 0745dd19-6b08-447a-8333-047aeaa1fa83/0745dd19-6b08-447a-8333-047aeaa1fa83.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.406056] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130050, 'name': CreateVM_Task, 'duration_secs': 0.574874} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.406281] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5d0fe42-a26d-41b2-a0b2-98e7d6e7154c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.407879] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.408444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.408615] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.408970] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.409563] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51c3bdc6-c1a4-454d-86f0-db96464c9372 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.413935] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 918.413935] env[61991]: value = "task-1130057" [ 918.413935] env[61991]: _type = "Task" [ 918.413935] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.415049] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 918.415049] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52790488-3653-55eb-6d02-ef47f054b46d" [ 918.415049] env[61991]: _type = "Task" [ 918.415049] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.426065] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130057, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.428772] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52790488-3653-55eb-6d02-ef47f054b46d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.470352] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-261d2df8-b27e-4aef-9664-39ffcd2cca24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.504235] env[61991]: DEBUG nova.network.neutron [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updated VIF entry in instance network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.504635] env[61991]: DEBUG nova.network.neutron [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.573877] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 918.634336] env[61991]: DEBUG oslo_vmware.api [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.444075} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.634466] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.634699] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.635169] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.635361] env[61991]: INFO nova.compute.manager [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Took 1.12 seconds to destroy the instance on the hypervisor. [ 918.635808] env[61991]: DEBUG oslo.service.loopingcall [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.636237] env[61991]: DEBUG nova.compute.manager [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.637028] env[61991]: DEBUG nova.network.neutron [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 918.682887] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130052, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.927617] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130057, 'name': Rename_Task, 'duration_secs': 0.141422} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.931396] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.931666] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52790488-3653-55eb-6d02-ef47f054b46d, 'name': SearchDatastore_Task, 'duration_secs': 0.025139} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.931874] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d71cff8-b922-4482-9757-585088fcf21e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.933980] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.934389] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.934648] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.934751] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.934932] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.935202] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad7697c3-0c3e-4714-8cac-d656ef955d65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.943990] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 918.943990] env[61991]: value = "task-1130058" [ 918.943990] env[61991]: _type = "Task" [ 918.943990] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.945583] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.945583] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.951849] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f190689-da20-4852-8297-887060711d9c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.964873] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.967287] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 918.967287] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c3a60e-a424-d2ae-e7d8-6ef668d70237" [ 918.967287] env[61991]: _type = "Task" [ 918.967287] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.979578] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c3a60e-a424-d2ae-e7d8-6ef668d70237, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.008264] env[61991]: DEBUG oslo_concurrency.lockutils [req-72bd444f-5b66-41fd-ba09-981786f09d84 req-611cc465-ede1-4eef-8cde-dc33a23064c1 service nova] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.099528] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.157297] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.184676] env[61991]: DEBUG nova.compute.manager [req-138cd184-047d-4006-ac2a-5f96630c224c req-63c40087-500e-47e2-80cd-41a8f14ecfbf service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Received event network-vif-deleted-4b76d924-ef23-455b-afc4-7ea48006e69e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.185146] env[61991]: INFO nova.compute.manager [req-138cd184-047d-4006-ac2a-5f96630c224c req-63c40087-500e-47e2-80cd-41a8f14ecfbf service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Neutron deleted interface 4b76d924-ef23-455b-afc4-7ea48006e69e; detaching it from the instance and deleting it from the info cache [ 919.185406] env[61991]: DEBUG nova.network.neutron [req-138cd184-047d-4006-ac2a-5f96630c224c req-63c40087-500e-47e2-80cd-41a8f14ecfbf service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.198840] env[61991]: DEBUG oslo_vmware.api [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130052, 'name': PowerOnVM_Task, 'duration_secs': 1.431972} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.199370] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.199563] env[61991]: INFO nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Took 9.34 seconds to spawn the instance on the hypervisor. [ 919.199806] env[61991]: DEBUG nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.202624] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.202758] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.202918] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.203342] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.203520] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.203587] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.203809] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.204073] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.204337] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.204537] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.204814] env[61991]: DEBUG nova.virt.hardware [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.205715] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de6aaf9-ba40-4858-8703-117403d11a6e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.210518] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538f41ed-4d3c-48d5-a0f3-796f5a0e68a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.221141] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.221376] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.235466] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba09721a-d52b-4644-ac13-f5c7837cc81f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.455508] env[61991]: DEBUG oslo_vmware.api [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130058, 'name': PowerOnVM_Task, 'duration_secs': 0.50585} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.459834] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.459834] env[61991]: INFO nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Took 7.05 seconds to spawn the instance on the hypervisor. [ 919.459834] env[61991]: DEBUG nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.460897] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0458bf98-c652-4570-9aa9-5f30cfce4fad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.482739] env[61991]: DEBUG nova.network.neutron [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.492142] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c3a60e-a424-d2ae-e7d8-6ef668d70237, 'name': SearchDatastore_Task, 'duration_secs': 0.016145} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.493897] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5924ec3-a7ed-47d2-90c8-ef0cf9af1439 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.499087] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dacf8d9-4218-42b1-85f2-29d80e289328 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.504727] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 919.504727] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524f33f4-7cdf-4d15-0df3-e22d151650b5" [ 919.504727] env[61991]: _type = "Task" [ 919.504727] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.517031] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87b7edd-65e2-4996-b4e3-4add6e807d36 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.526736] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524f33f4-7cdf-4d15-0df3-e22d151650b5, 'name': SearchDatastore_Task, 'duration_secs': 0.015077} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.527957] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.527957] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/5364f93c-ca53-480b-8397-c331dfdeed3f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.528491] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-331d9b5c-99ba-4995-9caa-90fde0f4e661 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.557324] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb60c3e9-7fb0-402c-8291-8a43de723fd9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.562178] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 919.562178] env[61991]: value = "task-1130059" [ 919.562178] env[61991]: _type = "Task" [ 919.562178] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.568675] env[61991]: DEBUG nova.network.neutron [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Successfully updated port: 7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.572331] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83015446-2d06-45eb-812b-22040b9cca90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.581140] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-9d1ac179-3d7e-4f0b-a80d-033eba447cab-b603ad92-a23f-4da6-b5dc-60d547ba32ae" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.581610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-9d1ac179-3d7e-4f0b-a80d-033eba447cab-b603ad92-a23f-4da6-b5dc-60d547ba32ae" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.583691] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130059, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.598530] env[61991]: DEBUG nova.compute.provider_tree [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.639812] env[61991]: DEBUG nova.compute.manager [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Received event network-vif-plugged-7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.640227] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] Acquiring lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.640606] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.640880] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.641150] env[61991]: DEBUG nova.compute.manager [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] No waiting events found dispatching network-vif-plugged-7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.641423] env[61991]: WARNING nova.compute.manager [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Received unexpected event network-vif-plugged-7cc61fcf-f52c-468a-8d9d-91635e8e25e9 for instance with vm_state building and task_state spawning. [ 919.641609] env[61991]: DEBUG nova.compute.manager [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Received event network-changed-7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.641865] env[61991]: DEBUG nova.compute.manager [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Refreshing instance network info cache due to event network-changed-7cc61fcf-f52c-468a-8d9d-91635e8e25e9. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.642107] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] Acquiring lock "refresh_cache-586bf51a-d1c4-44e8-863b-b229a3ca4f2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.642302] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] Acquired lock "refresh_cache-586bf51a-d1c4-44e8-863b-b229a3ca4f2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.642558] env[61991]: DEBUG nova.network.neutron [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Refreshing network info cache for port 7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.691353] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2cd8b61e-2594-4cd4-be63-b8cf98ae606e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.701263] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315f0028-6b10-43e2-b77a-5a5d329d314d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.738879] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.745773] env[61991]: DEBUG nova.compute.manager [req-138cd184-047d-4006-ac2a-5f96630c224c req-63c40087-500e-47e2-80cd-41a8f14ecfbf service nova] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Detach interface failed, port_id=4b76d924-ef23-455b-afc4-7ea48006e69e, reason: Instance 0dcf415e-b554-4c82-b127-9f0e22176031 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 919.751112] env[61991]: INFO nova.compute.manager [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Took 30.17 seconds to build instance. [ 919.983901] env[61991]: INFO nova.compute.manager [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Took 21.30 seconds to build instance. [ 919.987373] env[61991]: INFO nova.compute.manager [-] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Took 1.35 seconds to deallocate network for instance. [ 920.075169] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "refresh_cache-586bf51a-d1c4-44e8-863b-b229a3ca4f2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.075782] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130059, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.085483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.085607] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.086728] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9516e468-12f5-42e2-af4a-4e5176673ba8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.108683] env[61991]: DEBUG nova.scheduler.client.report [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.114111] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1739cba0-40aa-4b32-9dbb-b2185038fc66 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.144396] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Reconfiguring VM to detach interface {{(pid=61991) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 920.145402] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4365a26-6250-4e41-9f66-c7a32493d781 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.174832] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 920.174832] env[61991]: value = "task-1130060" [ 920.174832] env[61991]: _type = "Task" [ 920.174832] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.190443] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.237086] env[61991]: DEBUG nova.network.neutron [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 920.253966] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7861e844-a916-4dc7-8fc0-76ff7ca38087 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "077c5def-84cc-48f9-be90-410100666773" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.685s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.267767] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.346023] env[61991]: DEBUG nova.network.neutron [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.486568] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6d32f560-4332-4803-8307-bc2fc92aec5b tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.818s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.493168] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.574665] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130059, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653757} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.575429] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/5364f93c-ca53-480b-8397-c331dfdeed3f.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.575582] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.575829] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-733941cb-a1c3-4aeb-8b28-3adb74a1b3c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.582111] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 920.582111] env[61991]: value = "task-1130061" [ 920.582111] env[61991]: _type = "Task" [ 920.582111] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.590797] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130061, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.618458] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.619341] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.623360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.524s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.625053] env[61991]: INFO nova.compute.claims [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.687312] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.852920] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa8c0083-569c-4f4c-a529-9769857a906f req-4178ed15-a16d-48a2-8e20-e7e978fa4a78 service nova] Releasing lock "refresh_cache-586bf51a-d1c4-44e8-863b-b229a3ca4f2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.852920] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquired lock "refresh_cache-586bf51a-d1c4-44e8-863b-b229a3ca4f2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.852920] env[61991]: DEBUG nova.network.neutron [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 921.039261] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.039717] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.040038] env[61991]: INFO nova.compute.manager [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Shelving [ 921.092724] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130061, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070289} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.093481] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.093827] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a763eb-dbcf-475a-a230-f198f095e08f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.120858] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/5364f93c-ca53-480b-8397-c331dfdeed3f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.121562] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aba1f798-ded2-4e45-b821-6625f2e8b091 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.139908] env[61991]: DEBUG nova.compute.utils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.143322] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.143494] env[61991]: DEBUG nova.network.neutron [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 921.152043] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 921.152043] env[61991]: value = "task-1130062" [ 921.152043] env[61991]: _type = "Task" [ 921.152043] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.160763] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130062, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.189706] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.214344] env[61991]: DEBUG nova.policy [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e6810d3aced4a44bbfe9ba49270bc35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4a18b2a73f7643e1b89c9660a770e9da', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 921.420891] env[61991]: DEBUG nova.network.neutron [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.461864] env[61991]: DEBUG nova.compute.manager [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Received event network-changed-02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.462086] env[61991]: DEBUG nova.compute.manager [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Refreshing instance network info cache due to event network-changed-02083cbb-c0d3-414b-9882-6260c77dfd0c. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.462313] env[61991]: DEBUG oslo_concurrency.lockutils [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] Acquiring lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.462460] env[61991]: DEBUG oslo_concurrency.lockutils [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] Acquired lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.462623] env[61991]: DEBUG nova.network.neutron [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Refreshing network info cache for port 02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.551657] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.552060] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a117f596-9c5f-4522-8b6a-e7d937c527c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.556359] env[61991]: DEBUG nova.network.neutron [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Successfully created port: 2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.560877] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 921.560877] env[61991]: value = "task-1130063" [ 921.560877] env[61991]: _type = "Task" [ 921.560877] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.569707] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.592153] env[61991]: DEBUG nova.network.neutron [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Updating instance_info_cache with network_info: [{"id": "7cc61fcf-f52c-468a-8d9d-91635e8e25e9", "address": "fa:16:3e:e6:46:4d", "network": {"id": "5a18e9ae-c669-48ec-91da-c3ed168f5796", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-510110813-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06f21c8f25694d4d93dfd78bf10e539c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cc61fcf-f5", "ovs_interfaceid": "7cc61fcf-f52c-468a-8d9d-91635e8e25e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.644088] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.664751] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130062, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.691540] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.078020] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130063, 'name': PowerOffVM_Task, 'duration_secs': 0.291239} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.081092] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.082352] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6635eeb7-a3ca-4a12-b569-76d469204a76 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.104962] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Releasing lock "refresh_cache-586bf51a-d1c4-44e8-863b-b229a3ca4f2a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.105429] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Instance network_info: |[{"id": "7cc61fcf-f52c-468a-8d9d-91635e8e25e9", "address": "fa:16:3e:e6:46:4d", "network": {"id": "5a18e9ae-c669-48ec-91da-c3ed168f5796", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-510110813-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06f21c8f25694d4d93dfd78bf10e539c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7cc61fcf-f5", "ovs_interfaceid": "7cc61fcf-f52c-468a-8d9d-91635e8e25e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.106138] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:46:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ba6157eb-73cb-428a-9f46-99081165d7eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7cc61fcf-f52c-468a-8d9d-91635e8e25e9', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.114683] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Creating folder: Project (06f21c8f25694d4d93dfd78bf10e539c). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 922.115467] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e34318a-ac9c-4b30-812b-e35c9062ffd7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.119446] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e4f53bc-425b-4106-afcd-7e9373a81aca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.130537] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Created folder: Project (06f21c8f25694d4d93dfd78bf10e539c) in parent group-v246753. [ 922.131065] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Creating folder: Instances. Parent ref: group-v246929. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 922.131633] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c158b0c4-8b88-4708-b037-6466af7c0029 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.141798] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Created folder: Instances in parent group-v246929. [ 922.142218] env[61991]: DEBUG oslo.service.loopingcall [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.144150] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.144150] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-073cd4eb-3c1a-4d63-8306-ff4253ad3bd3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.174347] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.174347] env[61991]: value = "task-1130066" [ 922.174347] env[61991]: _type = "Task" [ 922.174347] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.174844] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130062, 'name': ReconfigVM_Task, 'duration_secs': 0.629042} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.175194] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/5364f93c-ca53-480b-8397-c331dfdeed3f.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.179803] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f010a6e-d7d7-4dad-8edd-f633d7b346a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.190362] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806e4c60-c448-47aa-8b20-7fca08945b1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.197759] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130066, 'name': CreateVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.198213] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 922.198213] env[61991]: value = "task-1130067" [ 922.198213] env[61991]: _type = "Task" [ 922.198213] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.208726] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fca480d-ada6-4ff1-a8e8-a4526ecc34da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.212725] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.219998] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130067, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.250973] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a3cd92-5334-469a-acc7-15f384b3a69d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.263888] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e67c26-ab13-41e4-8551-6a32dabbb6cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.281635] env[61991]: DEBUG nova.compute.provider_tree [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.348688] env[61991]: DEBUG nova.network.neutron [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updated VIF entry in instance network info cache for port 02083cbb-c0d3-414b-9882-6260c77dfd0c. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 922.349138] env[61991]: DEBUG nova.network.neutron [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updating instance_info_cache with network_info: [{"id": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "address": "fa:16:3e:2e:cc:6a", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02083cbb-c0", "ovs_interfaceid": "02083cbb-c0d3-414b-9882-6260c77dfd0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.633562] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 922.634406] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-283e2909-f4bb-4b1c-8b26-a4314cea70d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.642530] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 922.642530] env[61991]: value = "task-1130068" [ 922.642530] env[61991]: _type = "Task" [ 922.642530] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.650864] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130068, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.667449] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.684768] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130066, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.695856] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.698351] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.698629] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.698820] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.699063] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.699258] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.699451] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.699705] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.699913] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.700144] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.700362] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.700582] env[61991]: DEBUG nova.virt.hardware [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.701524] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666c78cc-e5dd-4164-9571-c4a9b5f3d4a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.717899] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130067, 'name': Rename_Task, 'duration_secs': 0.174251} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.718731] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.720292] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f3360b-0641-4182-9fe1-72305633ede7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.725337] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5f2eb5d-cd11-4680-aba1-faf99ea36214 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.740287] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 922.740287] env[61991]: value = "task-1130069" [ 922.740287] env[61991]: _type = "Task" [ 922.740287] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.749205] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130069, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.787160] env[61991]: DEBUG nova.scheduler.client.report [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.852183] env[61991]: DEBUG oslo_concurrency.lockutils [req-ea3197f1-f0ec-4b44-b036-d4091e2084a3 req-eee63c6a-af6c-4826-a56d-c48f22acad48 service nova] Releasing lock "refresh_cache-0745dd19-6b08-447a-8333-047aeaa1fa83" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.085084] env[61991]: DEBUG nova.compute.manager [req-a7eaebf5-d7d4-4675-bf8e-4d1326e08d7f req-54ff8c47-4821-40e6-86fa-746c22c7c9db service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Received event network-vif-plugged-2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.085452] env[61991]: DEBUG oslo_concurrency.lockutils [req-a7eaebf5-d7d4-4675-bf8e-4d1326e08d7f req-54ff8c47-4821-40e6-86fa-746c22c7c9db service nova] Acquiring lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.085790] env[61991]: DEBUG oslo_concurrency.lockutils [req-a7eaebf5-d7d4-4675-bf8e-4d1326e08d7f req-54ff8c47-4821-40e6-86fa-746c22c7c9db service nova] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.086063] env[61991]: DEBUG oslo_concurrency.lockutils [req-a7eaebf5-d7d4-4675-bf8e-4d1326e08d7f req-54ff8c47-4821-40e6-86fa-746c22c7c9db service nova] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.086306] env[61991]: DEBUG nova.compute.manager [req-a7eaebf5-d7d4-4675-bf8e-4d1326e08d7f req-54ff8c47-4821-40e6-86fa-746c22c7c9db service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] No waiting events found dispatching network-vif-plugged-2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.086524] env[61991]: WARNING nova.compute.manager [req-a7eaebf5-d7d4-4675-bf8e-4d1326e08d7f req-54ff8c47-4821-40e6-86fa-746c22c7c9db service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Received unexpected event network-vif-plugged-2dc3450f-efd8-4200-a182-988386f68d00 for instance with vm_state building and task_state spawning. [ 923.154287] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130068, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.188507] env[61991]: DEBUG nova.network.neutron [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Successfully updated port: 2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.203167] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130066, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.207368] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.251039] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130069, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.275043] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.275341] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.275532] env[61991]: DEBUG nova.compute.manager [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.276522] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d03cb75-583a-44ed-b5d3-3f4ea440311f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.283408] env[61991]: DEBUG nova.compute.manager [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 923.283943] env[61991]: DEBUG nova.objects.instance [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lazy-loading 'flavor' on Instance uuid 69a272a7-8da6-4fb2-95be-85cfb7d70d44 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.292337] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.669s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.292819] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 923.295330] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.028s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.296879] env[61991]: INFO nova.compute.claims [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.654835] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130068, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.693101] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.693101] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.693101] env[61991]: DEBUG nova.network.neutron [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.694127] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130066, 'name': CreateVM_Task, 'duration_secs': 1.461433} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.699688] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.700721] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.700908] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.701573] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.702770] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a00072d3-d083-41c3-938c-f144c53c8563 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.708978] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.711827] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 923.711827] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520f7a7a-4df5-72f0-4eca-b3330e2a6021" [ 923.711827] env[61991]: _type = "Task" [ 923.711827] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.509841] env[61991]: DEBUG nova.compute.utils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.515375] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.519117] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.519294] env[61991]: DEBUG nova.network.neutron [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 924.528142] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0b907ab-35fa-4298-98c7-4eba52c29276 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.531269] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.531505] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.538412] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520f7a7a-4df5-72f0-4eca-b3330e2a6021, 'name': SearchDatastore_Task, 'duration_secs': 0.012698} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.539139] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.539464] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.539744] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.539906] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.540130] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.549952] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1436820-a146-4d37-9afb-7485ffe13026 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.553773] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.554466] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130068, 'name': CreateSnapshot_Task, 'duration_secs': 1.098921} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.554705] env[61991]: DEBUG oslo_vmware.api [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130069, 'name': PowerOnVM_Task, 'duration_secs': 0.937195} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.556335] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 924.556629] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.556855] env[61991]: INFO nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Took 9.74 seconds to spawn the instance on the hypervisor. [ 924.557062] env[61991]: DEBUG nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.557418] env[61991]: DEBUG oslo_vmware.api [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 924.557418] env[61991]: value = "task-1130070" [ 924.557418] env[61991]: _type = "Task" [ 924.557418] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.558516] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac445ad-6051-4f00-9861-873c3eb3798b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.562363] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7d4267-cabc-4b44-91e1-51c0eb6aa8d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.569805] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.570107] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.572065] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eab68dfa-5f11-439e-be5d-7b282c2052e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.575282] env[61991]: DEBUG nova.network.neutron [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.592012] env[61991]: DEBUG oslo_vmware.api [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.596019] env[61991]: DEBUG nova.policy [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6422dfa4c424fbaab78c3e72a0e6669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dddb51029854105bc6c9c8724181d39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.596712] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 924.596712] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529419e4-4b8f-1a4a-3517-e841cc7234ff" [ 924.596712] env[61991]: _type = "Task" [ 924.596712] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.605812] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529419e4-4b8f-1a4a-3517-e841cc7234ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.824531] env[61991]: DEBUG nova.network.neutron [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance_info_cache with network_info: [{"id": "2dc3450f-efd8-4200-a182-988386f68d00", "address": "fa:16:3e:de:38:21", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dc3450f-ef", "ovs_interfaceid": "2dc3450f-efd8-4200-a182-988386f68d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.875493] env[61991]: DEBUG nova.network.neutron [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Successfully created port: a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.020509] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.028980] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.040266] env[61991]: DEBUG nova.compute.utils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.078927] env[61991]: DEBUG oslo_vmware.api [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130070, 'name': PowerOffVM_Task, 'duration_secs': 0.255479} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.079219] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.079395] env[61991]: DEBUG nova.compute.manager [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.080589] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95938f53-9d3f-4eea-9c91-b81ab1bbf601 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.107612] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 925.108179] env[61991]: INFO nova.compute.manager [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Took 25.77 seconds to build instance. [ 925.111100] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5095a102-ec6d-4211-996b-981f0f4a6ef3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.117660] env[61991]: DEBUG nova.compute.manager [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Received event network-changed-2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.117872] env[61991]: DEBUG nova.compute.manager [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Refreshing instance network info cache due to event network-changed-2dc3450f-efd8-4200-a182-988386f68d00. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.118136] env[61991]: DEBUG oslo_concurrency.lockutils [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] Acquiring lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.122509] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529419e4-4b8f-1a4a-3517-e841cc7234ff, 'name': SearchDatastore_Task, 'duration_secs': 0.018687} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.126513] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 925.126513] env[61991]: value = "task-1130071" [ 925.126513] env[61991]: _type = "Task" [ 925.126513] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.127871] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1972f60d-a028-4dd9-8adf-f4ea53ea60a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.136933] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 925.136933] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522c7098-bd82-30ce-46a6-950e9095954b" [ 925.136933] env[61991]: _type = "Task" [ 925.136933] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.139948] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130071, 'name': CloneVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.150698] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522c7098-bd82-30ce-46a6-950e9095954b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.324270] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32b22ba-cdab-46d8-8893-1b3b620e74be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.327892] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.327892] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Instance network_info: |[{"id": "2dc3450f-efd8-4200-a182-988386f68d00", "address": "fa:16:3e:de:38:21", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dc3450f-ef", "ovs_interfaceid": "2dc3450f-efd8-4200-a182-988386f68d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 925.328336] env[61991]: DEBUG oslo_concurrency.lockutils [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] Acquired lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.328510] env[61991]: DEBUG nova.network.neutron [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Refreshing network info cache for port 2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.330542] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:38:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2dc3450f-efd8-4200-a182-988386f68d00', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.337866] env[61991]: DEBUG oslo.service.loopingcall [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.339490] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.339726] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-491313e8-d773-4dd4-9f36-45a9a6f47b10 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.361529] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defb9cee-08a6-4e90-80ac-c3a015d7d47f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.366648] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.366648] env[61991]: value = "task-1130072" [ 925.366648] env[61991]: _type = "Task" [ 925.366648] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.399408] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d3c859-35b8-4f68-a497-401478706591 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.405308] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130072, 'name': CreateVM_Task} progress is 15%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.410617] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b29317-0336-44ab-8646-1d9886d0c214 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.425656] env[61991]: DEBUG nova.compute.provider_tree [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.528212] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.545408] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.611046] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f8fa031a-29f2-4388-a911-2b121132a411 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.336s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.616084] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c7e44e5-dfcd-4abb-8f5f-0afd910f23d9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.288s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.642231] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130071, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.652622] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522c7098-bd82-30ce-46a6-950e9095954b, 'name': SearchDatastore_Task, 'duration_secs': 0.018642} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.652954] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.653290] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 586bf51a-d1c4-44e8-863b-b229a3ca4f2a/586bf51a-d1c4-44e8-863b-b229a3ca4f2a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 925.653589] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8926fe6-52c9-4cae-83e4-c73f4acde14e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.661356] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 925.661356] env[61991]: value = "task-1130073" [ 925.661356] env[61991]: _type = "Task" [ 925.661356] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.671469] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.878656] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130072, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.929132] env[61991]: DEBUG nova.scheduler.client.report [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 926.028683] env[61991]: DEBUG oslo_vmware.api [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130060, 'name': ReconfigVM_Task, 'duration_secs': 5.826637} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.032859] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.035373] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.035825] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Reconfigured VM to detach interface {{(pid=61991) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 926.079547] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.079767] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.079887] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.080081] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.080235] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.080684] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.080684] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.080903] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.080966] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.081138] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.081341] env[61991]: DEBUG nova.virt.hardware [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.083266] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e96ade-14e0-4139-88a7-32441ad4b3b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.092482] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06074dcb-131d-459b-8dc4-831e23cb4fdc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.144590] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130071, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.173783] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130073, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.313640] env[61991]: DEBUG nova.network.neutron [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updated VIF entry in instance network info cache for port 2dc3450f-efd8-4200-a182-988386f68d00. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 926.313640] env[61991]: DEBUG nova.network.neutron [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance_info_cache with network_info: [{"id": "2dc3450f-efd8-4200-a182-988386f68d00", "address": "fa:16:3e:de:38:21", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dc3450f-ef", "ovs_interfaceid": "2dc3450f-efd8-4200-a182-988386f68d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.320801] env[61991]: DEBUG nova.objects.instance [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lazy-loading 'flavor' on Instance uuid 69a272a7-8da6-4fb2-95be-85cfb7d70d44 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.377601] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130072, 'name': CreateVM_Task, 'duration_secs': 0.6361} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.377793] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 926.378508] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.378677] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.379132] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.379271] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4c1a6b8-3d40-400b-bcee-aef0d38a7acb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.392830] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 926.392830] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5215b9d7-0a70-1824-9ab0-0972c52f2b2f" [ 926.392830] env[61991]: _type = "Task" [ 926.392830] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.401649] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5215b9d7-0a70-1824-9ab0-0972c52f2b2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.416607] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.419958] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.419958] env[61991]: INFO nova.compute.manager [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Attaching volume e1d09948-3670-4f76-ad45-f10a94df1f32 to /dev/sdb [ 926.439597] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.144s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.440310] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.443019] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.950s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.443322] env[61991]: DEBUG nova.objects.instance [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lazy-loading 'resources' on Instance uuid 0dcf415e-b554-4c82-b127-9f0e22176031 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.459668] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e1e860-83be-4b65-bdfc-6cf24314aa78 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.468065] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb636469-08fc-4499-baf6-cdc70b1258d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.484120] env[61991]: DEBUG nova.virt.block_device [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating existing volume attachment record: c6020674-b6db-4709-b6b9-44c0fdd51beb {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 926.556045] env[61991]: DEBUG nova.compute.manager [req-c5bc776e-e251-44cf-8a43-f76793a32860 req-49ed5db4-47f4-42fb-9faf-d611199a312c service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Received event network-vif-plugged-a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.556599] env[61991]: DEBUG oslo_concurrency.lockutils [req-c5bc776e-e251-44cf-8a43-f76793a32860 req-49ed5db4-47f4-42fb-9faf-d611199a312c service nova] Acquiring lock "68268135-4cf4-4732-9104-1720bbb21acf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.556872] env[61991]: DEBUG oslo_concurrency.lockutils [req-c5bc776e-e251-44cf-8a43-f76793a32860 req-49ed5db4-47f4-42fb-9faf-d611199a312c service nova] Lock "68268135-4cf4-4732-9104-1720bbb21acf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.557076] env[61991]: DEBUG oslo_concurrency.lockutils [req-c5bc776e-e251-44cf-8a43-f76793a32860 req-49ed5db4-47f4-42fb-9faf-d611199a312c service nova] Lock "68268135-4cf4-4732-9104-1720bbb21acf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.557252] env[61991]: DEBUG nova.compute.manager [req-c5bc776e-e251-44cf-8a43-f76793a32860 req-49ed5db4-47f4-42fb-9faf-d611199a312c service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] No waiting events found dispatching network-vif-plugged-a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.557545] env[61991]: WARNING nova.compute.manager [req-c5bc776e-e251-44cf-8a43-f76793a32860 req-49ed5db4-47f4-42fb-9faf-d611199a312c service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Received unexpected event network-vif-plugged-a511534a-b3ea-4087-81d0-c1b3277ecb0f for instance with vm_state building and task_state spawning. [ 926.645939] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130071, 'name': CloneVM_Task, 'duration_secs': 1.204082} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.647587] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Created linked-clone VM from snapshot [ 926.648467] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3361d425-9642-40cf-b01a-ea5f78fa9d65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.657417] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Uploading image 5cf3221b-0c05-4b11-a46c-192b39b794ad {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 926.673218] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130073, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565807} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.673419] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 586bf51a-d1c4-44e8-863b-b229a3ca4f2a/586bf51a-d1c4-44e8-863b-b229a3ca4f2a.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 926.673667] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.673936] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-838d37e3-e613-4557-b9df-db91a154d1a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.679968] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 926.679968] env[61991]: value = "task-1130075" [ 926.679968] env[61991]: _type = "Task" [ 926.679968] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.686951] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 926.686951] env[61991]: value = "vm-246933" [ 926.686951] env[61991]: _type = "VirtualMachine" [ 926.686951] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 926.687315] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a7774ee8-1df9-4a4b-a912-388d144265bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.693665] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130075, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.695108] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lease: (returnval){ [ 926.695108] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527ff4d3-114c-3870-3adf-d959524a3ff9" [ 926.695108] env[61991]: _type = "HttpNfcLease" [ 926.695108] env[61991]: } obtained for exporting VM: (result){ [ 926.695108] env[61991]: value = "vm-246933" [ 926.695108] env[61991]: _type = "VirtualMachine" [ 926.695108] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 926.696520] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the lease: (returnval){ [ 926.696520] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527ff4d3-114c-3870-3adf-d959524a3ff9" [ 926.696520] env[61991]: _type = "HttpNfcLease" [ 926.696520] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 926.701671] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 926.701671] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527ff4d3-114c-3870-3adf-d959524a3ff9" [ 926.701671] env[61991]: _type = "HttpNfcLease" [ 926.701671] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 926.788432] env[61991]: DEBUG nova.network.neutron [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Successfully updated port: a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.817483] env[61991]: DEBUG oslo_concurrency.lockutils [req-ac5f1e76-cb06-42c7-a738-eb87ccd599f6 req-0acfb3d6-c22f-4708-8e1c-8a1cf1389996 service nova] Releasing lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.826023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.826276] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquired lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.826473] env[61991]: DEBUG nova.network.neutron [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.826669] env[61991]: DEBUG nova.objects.instance [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lazy-loading 'info_cache' on Instance uuid 69a272a7-8da6-4fb2-95be-85cfb7d70d44 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.904025] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5215b9d7-0a70-1824-9ab0-0972c52f2b2f, 'name': SearchDatastore_Task, 'duration_secs': 0.026581} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.904419] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.904466] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.904669] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.904820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.905008] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.905278] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86df748c-1f8d-49b9-9246-aeaf4513ad00 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.916355] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.916553] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.917709] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0208d18-b96c-487f-a36b-9c7d4f107e8d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.923808] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 926.923808] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5233db9b-6d95-b52d-e1ee-0574c5ea0291" [ 926.923808] env[61991]: _type = "Task" [ 926.923808] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.932618] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5233db9b-6d95-b52d-e1ee-0574c5ea0291, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.947070] env[61991]: DEBUG nova.compute.utils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.950406] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.950581] env[61991]: DEBUG nova.network.neutron [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 927.003812] env[61991]: DEBUG nova.policy [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c5daf21b21b40c9a54bed64a64538bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '822d7e3c678e4defa24bb4d8439a62de', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.191997] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130075, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101874} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.195183] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.196338] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0376a1f8-387a-42d7-bace-985fd99d22e7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.206698] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 927.206698] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527ff4d3-114c-3870-3adf-d959524a3ff9" [ 927.206698] env[61991]: _type = "HttpNfcLease" [ 927.206698] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 927.216516] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 927.216516] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527ff4d3-114c-3870-3adf-d959524a3ff9" [ 927.216516] env[61991]: _type = "HttpNfcLease" [ 927.216516] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 927.225465] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 586bf51a-d1c4-44e8-863b-b229a3ca4f2a/586bf51a-d1c4-44e8-863b-b229a3ca4f2a.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.229078] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76195115-7992-4640-9c73-54237fc94b4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.232372] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b7bd081-2a4b-4ccb-b7ab-ca2322a000eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.256239] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52289603-565b-2917-c64e-131de310bc9f/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 927.256452] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52289603-565b-2917-c64e-131de310bc9f/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 927.259474] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 927.259474] env[61991]: value = "task-1130079" [ 927.259474] env[61991]: _type = "Task" [ 927.259474] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.330244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.330385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.330576] env[61991]: DEBUG nova.network.neutron [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.335759] env[61991]: DEBUG nova.objects.base [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Object Instance<69a272a7-8da6-4fb2-95be-85cfb7d70d44> lazy-loaded attributes: flavor,info_cache {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 927.345507] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130079, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.356212] env[61991]: DEBUG nova.network.neutron [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Successfully created port: 4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.362470] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a97a321-a8e3-442d-bf3a-addca7230ae3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.373908] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9f6c2d-dfba-4aa8-bbb2-594921a77b20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.410616] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9c6e7f-5382-48b1-9ec7-349c85a8cc71 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.421145] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff80db9e-8054-4f9c-9a00-8015f1c82c2a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.441159] env[61991]: DEBUG nova.compute.provider_tree [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.447467] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5233db9b-6d95-b52d-e1ee-0574c5ea0291, 'name': SearchDatastore_Task, 'duration_secs': 0.026205} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.448737] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd0f4d9d-cf95-4f85-831d-7335fb7133f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.453068] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.458939] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-781f78ba-a450-440a-8b9a-3e65c13b731d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.463392] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 927.463392] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a118d0-e3fd-4e01-883b-aeb5580f679d" [ 927.463392] env[61991]: _type = "Task" [ 927.463392] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.477062] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a118d0-e3fd-4e01-883b-aeb5580f679d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.649156] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.649581] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.649823] env[61991]: DEBUG nova.network.neutron [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.776846] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130079, 'name': ReconfigVM_Task, 'duration_secs': 0.482929} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.776846] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 586bf51a-d1c4-44e8-863b-b229a3ca4f2a/586bf51a-d1c4-44e8-863b-b229a3ca4f2a.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.777072] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5912efca-8c34-4dad-bdde-a82e5dcc5887 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.786152] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 927.786152] env[61991]: value = "task-1130080" [ 927.786152] env[61991]: _type = "Task" [ 927.786152] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.800847] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130080, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.928364] env[61991]: DEBUG nova.network.neutron [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.945641] env[61991]: DEBUG nova.scheduler.client.report [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.978332] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a118d0-e3fd-4e01-883b-aeb5580f679d, 'name': SearchDatastore_Task, 'duration_secs': 0.019979} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.978973] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.979403] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514/3b647920-4a69-4d1d-adb1-6fbf716e5514.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 927.979978] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f30a185-e2af-4fec-ac24-f3992c33a347 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.992874] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 927.992874] env[61991]: value = "task-1130081" [ 927.992874] env[61991]: _type = "Task" [ 927.992874] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.013160] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.306732] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130080, 'name': Rename_Task, 'duration_secs': 0.185005} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.307109] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.307692] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58c21644-c206-49f8-a515-39b26e296845 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.317356] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 928.317356] env[61991]: value = "task-1130082" [ 928.317356] env[61991]: _type = "Task" [ 928.317356] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.327110] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130082, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.428361] env[61991]: INFO nova.network.neutron [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Port b603ad92-a23f-4da6-b5dc-60d547ba32ae from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 928.433025] env[61991]: DEBUG nova.network.neutron [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.454678] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.011s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.475694] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.486010] env[61991]: INFO nova.scheduler.client.report [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocations for instance 0dcf415e-b554-4c82-b127-9f0e22176031 [ 928.508186] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130081, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.570017] env[61991]: DEBUG nova.network.neutron [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance_info_cache with network_info: [{"id": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "address": "fa:16:3e:d7:45:df", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa511534a-b3", "ovs_interfaceid": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.644917] env[61991]: DEBUG nova.network.neutron [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Updating instance_info_cache with network_info: [{"id": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "address": "fa:16:3e:96:ee:74", "network": {"id": "8360e59c-e7d0-4f6a-8355-5d13354ed3b8", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1572467768-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c696231bd014ee5a67e4c3719f94ab4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2f5e5e2-e460-49ce-aa24-232e4a8007af", "external-id": "nsx-vlan-transportzone-503", "segmentation_id": 503, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9cd766-2c", "ovs_interfaceid": "9b9cd766-2c3d-4f0a-beb3-30531d95c2ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.828976] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130082, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.934963] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.995244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a8837439-8cc9-4860-985c-441d13630ca6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "0dcf415e-b554-4c82-b127-9f0e22176031" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.482s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.010062] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673841} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.010452] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514/3b647920-4a69-4d1d-adb1-6fbf716e5514.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.010908] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.012357] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f2817f3-6e6b-43dc-871e-893f90b74019 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.020999] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 929.020999] env[61991]: value = "task-1130084" [ 929.020999] env[61991]: _type = "Task" [ 929.020999] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.031513] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130084, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.051403] env[61991]: DEBUG nova.compute.manager [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Received event network-changed-a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.051871] env[61991]: DEBUG nova.compute.manager [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Refreshing instance network info cache due to event network-changed-a511534a-b3ea-4087-81d0-c1b3277ecb0f. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.052286] env[61991]: DEBUG oslo_concurrency.lockutils [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] Acquiring lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.070182] env[61991]: DEBUG nova.network.neutron [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Successfully updated port: 4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.077207] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.077638] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Instance network_info: |[{"id": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "address": "fa:16:3e:d7:45:df", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa511534a-b3", "ovs_interfaceid": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 929.078177] env[61991]: DEBUG oslo_concurrency.lockutils [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] Acquired lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.078493] env[61991]: DEBUG nova.network.neutron [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Refreshing network info cache for port a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 929.080839] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:45:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a511534a-b3ea-4087-81d0-c1b3277ecb0f', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.090169] env[61991]: DEBUG oslo.service.loopingcall [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.091938] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.092248] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c7841e6-253c-4c68-a912-94eb1d2c95c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.119717] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.119717] env[61991]: value = "task-1130085" [ 929.119717] env[61991]: _type = "Task" [ 929.119717] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.132729] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130085, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.148103] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Releasing lock "refresh_cache-69a272a7-8da6-4fb2-95be-85cfb7d70d44" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.331132] env[61991]: DEBUG oslo_vmware.api [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130082, 'name': PowerOnVM_Task, 'duration_secs': 0.615071} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.332888] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.332888] env[61991]: INFO nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Took 10.17 seconds to spawn the instance on the hypervisor. [ 929.332888] env[61991]: DEBUG nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.334253] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38cd922f-afef-444a-ad78-0abd29f2718a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.440059] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e56a7c8-5a29-4939-99ce-3cb2a8d16b8f tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-9d1ac179-3d7e-4f0b-a80d-033eba447cab-b603ad92-a23f-4da6-b5dc-60d547ba32ae" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.858s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.537707] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130084, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074168} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.537707] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.537707] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a84613-c94c-41b1-9105-3e8b9adadc29 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.563265] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514/3b647920-4a69-4d1d-adb1-6fbf716e5514.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.563265] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01d55e6e-fd41-4813-a200-9b662cc9718f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.582583] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.582583] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.582583] env[61991]: DEBUG nova.network.neutron [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.597399] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 929.597399] env[61991]: value = "task-1130086" [ 929.597399] env[61991]: _type = "Task" [ 929.597399] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.610880] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130086, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.634928] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130085, 'name': CreateVM_Task, 'duration_secs': 0.514517} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.637338] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.638522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.638700] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.639039] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 929.639611] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c820c5c-46f6-42a1-b998-5fa178df7a9a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.645152] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 929.645152] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e6d8ae-2a9f-91c7-f843-59e35511e883" [ 929.645152] env[61991]: _type = "Task" [ 929.645152] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.651780] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.652642] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec6d7fc7-bf76-4b3b-aae3-fe84f4e4bfe3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.656939] env[61991]: DEBUG nova.compute.manager [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.657318] env[61991]: DEBUG nova.compute.manager [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing instance network info cache due to event network-changed-f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.657389] env[61991]: DEBUG oslo_concurrency.lockutils [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] Acquiring lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.657526] env[61991]: DEBUG oslo_concurrency.lockutils [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] Acquired lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.657687] env[61991]: DEBUG nova.network.neutron [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Refreshing network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 929.663127] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e6d8ae-2a9f-91c7-f843-59e35511e883, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.669371] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.669578] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.669752] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.669947] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.670292] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.670371] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.670634] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.670810] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.670985] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.671168] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.671447] env[61991]: DEBUG nova.virt.hardware [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.675371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63498b76-1908-46e8-8c88-a7fce552d173 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.680969] env[61991]: DEBUG oslo_vmware.api [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 929.680969] env[61991]: value = "task-1130087" [ 929.680969] env[61991]: _type = "Task" [ 929.680969] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.691595] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f492ab-3240-a946-852b-d64628acbe51/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 929.692775] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0932c3bc-23f0-45b6-babb-7408ecdfc132 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.696757] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f81ccab-a940-435c-a97d-9ca4e86f1865 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.705416] env[61991]: DEBUG oslo_vmware.api [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130087, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.713401] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f492ab-3240-a946-852b-d64628acbe51/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 929.713576] env[61991]: ERROR oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f492ab-3240-a946-852b-d64628acbe51/disk-0.vmdk due to incomplete transfer. [ 929.721806] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d798b9b7-6c42-42b7-b037-d43068e3157f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.732071] env[61991]: DEBUG oslo_vmware.rw_handles [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f492ab-3240-a946-852b-d64628acbe51/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 929.732532] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Uploaded image 9a5d17c0-8aa4-41d4-8548-25843981601a to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 929.735303] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 929.736056] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7f4ff4f2-2ef1-41ab-a52f-1f005f591eb1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.744795] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 929.744795] env[61991]: value = "task-1130088" [ 929.744795] env[61991]: _type = "Task" [ 929.744795] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.754479] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130088, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.856595] env[61991]: INFO nova.compute.manager [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Took 24.82 seconds to build instance. [ 929.953687] env[61991]: DEBUG nova.network.neutron [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updated VIF entry in instance network info cache for port a511534a-b3ea-4087-81d0-c1b3277ecb0f. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 929.954219] env[61991]: DEBUG nova.network.neutron [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance_info_cache with network_info: [{"id": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "address": "fa:16:3e:d7:45:df", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa511534a-b3", "ovs_interfaceid": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.109310] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130086, 'name': ReconfigVM_Task, 'duration_secs': 0.434228} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.109658] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514/3b647920-4a69-4d1d-adb1-6fbf716e5514.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.110398] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83da280a-b394-400a-9a71-44e84f6b5b72 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.119857] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 930.119857] env[61991]: value = "task-1130089" [ 930.119857] env[61991]: _type = "Task" [ 930.119857] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.126453] env[61991]: DEBUG nova.network.neutron [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 930.132685] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130089, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.159419] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e6d8ae-2a9f-91c7-f843-59e35511e883, 'name': SearchDatastore_Task, 'duration_secs': 0.017894} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.160178] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.160597] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.161042] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.161340] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.161696] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.162147] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0216463-6f1e-44ee-b2e1-e721ab05e7a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.174943] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.175487] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.176488] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2c9ed81-8ee3-48c2-8eed-31f2027b6365 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.183167] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 930.183167] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527c2ecc-3fdf-6e94-feea-664fc9092edb" [ 930.183167] env[61991]: _type = "Task" [ 930.183167] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.205023] env[61991]: DEBUG oslo_vmware.api [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130087, 'name': PowerOnVM_Task, 'duration_secs': 0.467727} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.205023] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527c2ecc-3fdf-6e94-feea-664fc9092edb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.205023] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 930.205023] env[61991]: DEBUG nova.compute.manager [None req-31ff3916-6a0f-4cc2-9f01-514a9b48c245 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.205966] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49b651b-502e-44c9-a430-a496166d9c2d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.259107] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130088, 'name': Destroy_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.341189] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-902f158c-fcc4-40a2-abbd-491bdfd4b77a-b603ad92-a23f-4da6-b5dc-60d547ba32ae" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.341189] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-902f158c-fcc4-40a2-abbd-491bdfd4b77a-b603ad92-a23f-4da6-b5dc-60d547ba32ae" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.341189] env[61991]: DEBUG nova.objects.instance [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'flavor' on Instance uuid 902f158c-fcc4-40a2-abbd-491bdfd4b77a {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.345903] env[61991]: DEBUG nova.network.neutron [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.358942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-83ec3bf7-652f-4aae-8bba-47b3f8637516 tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.334s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.420373] env[61991]: DEBUG nova.network.neutron [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updated VIF entry in instance network info cache for port f07a2552-10f4-48c4-9a79-b66506f34302. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 930.420770] env[61991]: DEBUG nova.network.neutron [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [{"id": "f07a2552-10f4-48c4-9a79-b66506f34302", "address": "fa:16:3e:69:73:4f", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07a2552-10", "ovs_interfaceid": "f07a2552-10f4-48c4-9a79-b66506f34302", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.457326] env[61991]: DEBUG oslo_concurrency.lockutils [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] Releasing lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.457700] env[61991]: DEBUG nova.compute.manager [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.457895] env[61991]: DEBUG nova.compute.manager [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing instance network info cache due to event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.458167] env[61991]: DEBUG oslo_concurrency.lockutils [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.458344] env[61991]: DEBUG oslo_concurrency.lockutils [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.458518] env[61991]: DEBUG nova.network.neutron [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.630571] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130089, 'name': Rename_Task, 'duration_secs': 0.232172} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.630865] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.631149] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-020a77f4-e46b-4312-b447-f0aa193d71aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.638745] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 930.638745] env[61991]: value = "task-1130090" [ 930.638745] env[61991]: _type = "Task" [ 930.638745] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.647762] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.694309] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527c2ecc-3fdf-6e94-feea-664fc9092edb, 'name': SearchDatastore_Task, 'duration_secs': 0.016433} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.695128] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0efe474-14e8-4f57-a0cb-ee9cff9b53ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.701608] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 930.701608] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52794f6e-4af1-1cd2-46bc-2bf876379da0" [ 930.701608] env[61991]: _type = "Task" [ 930.701608] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.711052] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52794f6e-4af1-1cd2-46bc-2bf876379da0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.757789] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130088, 'name': Destroy_Task, 'duration_secs': 0.544151} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.758148] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Destroyed the VM [ 930.758441] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 930.758727] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0e136b90-a77b-4c2a-8080-f563895640c9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.767276] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 930.767276] env[61991]: value = "task-1130091" [ 930.767276] env[61991]: _type = "Task" [ 930.767276] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.778268] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130091, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.848806] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.849292] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Instance network_info: |[{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.853145] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:7e:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e506178-06a6-49c6-a6b3-fa24ebafc988', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.861132] env[61991]: DEBUG oslo.service.loopingcall [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.861421] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.862634] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ee78792-62ea-4d3b-b0bc-e72f655a3197 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.885576] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.885576] env[61991]: value = "task-1130092" [ 930.885576] env[61991]: _type = "Task" [ 930.885576] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.898298] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130092, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.926057] env[61991]: DEBUG oslo_concurrency.lockutils [req-e10d873c-bc99-4e36-ae1a-1ceaf537f710 req-a48a42f8-7972-4e51-b3cb-ff2502682811 service nova] Releasing lock "refresh_cache-9d1ac179-3d7e-4f0b-a80d-033eba447cab" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.968664] env[61991]: DEBUG nova.objects.instance [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'pci_requests' on Instance uuid 902f158c-fcc4-40a2-abbd-491bdfd4b77a {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.039704] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 931.040060] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 931.041151] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf207fde-075b-451c-8786-6f591865c961 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.060810] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f289c0aa-e0cc-4fa0-afc2-fc67ff308ead {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.089012] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] volume-e1d09948-3670-4f76-ad45-f10a94df1f32/volume-e1d09948-3670-4f76-ad45-f10a94df1f32.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.089480] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e0d22f8-b793-42ad-9d74-e55e97802b0c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.111572] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 931.111572] env[61991]: value = "task-1130093" [ 931.111572] env[61991]: _type = "Task" [ 931.111572] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.122227] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130093, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.150232] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130090, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.212783] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52794f6e-4af1-1cd2-46bc-2bf876379da0, 'name': SearchDatastore_Task, 'duration_secs': 0.01834} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.213075] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.213361] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf/68268135-4cf4-4732-9104-1720bbb21acf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.213641] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5290942e-4874-4194-ba34-13f2e207412e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.221484] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 931.221484] env[61991]: value = "task-1130094" [ 931.221484] env[61991]: _type = "Task" [ 931.221484] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.230864] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.281011] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130091, 'name': RemoveSnapshot_Task} progress is 50%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.400287] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130092, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.472589] env[61991]: DEBUG nova.objects.base [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Object Instance<902f158c-fcc4-40a2-abbd-491bdfd4b77a> lazy-loaded attributes: flavor,pci_requests {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 931.472903] env[61991]: DEBUG nova.network.neutron [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.521442] env[61991]: DEBUG nova.network.neutron [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updated VIF entry in instance network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.522023] env[61991]: DEBUG nova.network.neutron [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.555124] env[61991]: DEBUG nova.policy [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ad1a14facd9469bb9a53317f853c5b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd58eaa6ccc4f46e3a7606f55605d8e15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 931.624786] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130093, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.652795] env[61991]: DEBUG oslo_vmware.api [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130090, 'name': PowerOnVM_Task, 'duration_secs': 0.756251} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.653248] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.653405] env[61991]: INFO nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Took 8.99 seconds to spawn the instance on the hypervisor. [ 931.653636] env[61991]: DEBUG nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.654774] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a625b0c5-83ec-4c86-89ca-3fee17bcebd3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.672101] env[61991]: DEBUG nova.compute.manager [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Received event network-vif-plugged-4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.672101] env[61991]: DEBUG oslo_concurrency.lockutils [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.672393] env[61991]: DEBUG oslo_concurrency.lockutils [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.673127] env[61991]: DEBUG oslo_concurrency.lockutils [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.673237] env[61991]: DEBUG nova.compute.manager [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] No waiting events found dispatching network-vif-plugged-4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.673459] env[61991]: WARNING nova.compute.manager [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Received unexpected event network-vif-plugged-4e506178-06a6-49c6-a6b3-fa24ebafc988 for instance with vm_state building and task_state spawning. [ 931.673657] env[61991]: DEBUG nova.compute.manager [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Received event network-changed-4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.673823] env[61991]: DEBUG nova.compute.manager [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Refreshing instance network info cache due to event network-changed-4e506178-06a6-49c6-a6b3-fa24ebafc988. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.674039] env[61991]: DEBUG oslo_concurrency.lockutils [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.674198] env[61991]: DEBUG oslo_concurrency.lockutils [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.674377] env[61991]: DEBUG nova.network.neutron [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Refreshing network info cache for port 4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 931.736302] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130094, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.780040] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130091, 'name': RemoveSnapshot_Task} progress is 50%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.902545] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130092, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.025651] env[61991]: DEBUG oslo_concurrency.lockutils [req-f81bc8b6-ad30-4419-91ce-c40f7ccccf6d req-223bcafa-0256-48fb-9e60-2344400a88d6 service nova] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.093197] env[61991]: DEBUG nova.compute.manager [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.093350] env[61991]: DEBUG nova.compute.manager [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing instance network info cache due to event network-changed-e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.093609] env[61991]: DEBUG oslo_concurrency.lockutils [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.093821] env[61991]: DEBUG oslo_concurrency.lockutils [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.093992] env[61991]: DEBUG nova.network.neutron [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.122810] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130093, 'name': ReconfigVM_Task, 'duration_secs': 0.730682} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.123115] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfigured VM instance instance-00000054 to attach disk [datastore1] volume-e1d09948-3670-4f76-ad45-f10a94df1f32/volume-e1d09948-3670-4f76-ad45-f10a94df1f32.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.128144] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-945b7aec-cf57-49f4-a764-552a6aaad2f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.144776] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 932.144776] env[61991]: value = "task-1130095" [ 932.144776] env[61991]: _type = "Task" [ 932.144776] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.154358] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130095, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.180389] env[61991]: INFO nova.compute.manager [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Took 14.74 seconds to build instance. [ 932.233709] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63902} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.233986] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf/68268135-4cf4-4732-9104-1720bbb21acf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.234267] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.234528] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33086092-633d-4943-9284-003c33c71cfb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.242537] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 932.242537] env[61991]: value = "task-1130096" [ 932.242537] env[61991]: _type = "Task" [ 932.242537] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.253551] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130096, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.279439] env[61991]: DEBUG oslo_vmware.api [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130091, 'name': RemoveSnapshot_Task, 'duration_secs': 1.149904} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.279849] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 932.280137] env[61991]: INFO nova.compute.manager [None req-c92f55a5-6d0a-4e15-b8f1-ad0ea67c1936 tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Took 18.68 seconds to snapshot the instance on the hypervisor. [ 932.399661] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130092, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.536532] env[61991]: DEBUG nova.network.neutron [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updated VIF entry in instance network info cache for port 4e506178-06a6-49c6-a6b3-fa24ebafc988. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 932.536956] env[61991]: DEBUG nova.network.neutron [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.654896] env[61991]: DEBUG oslo_vmware.api [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130095, 'name': ReconfigVM_Task, 'duration_secs': 0.166405} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.655245] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 932.682839] env[61991]: DEBUG oslo_concurrency.lockutils [None req-057fe455-485e-4725-ab0b-a50c8755b2cf tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.249s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.752908] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130096, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102617} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.753251] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.754062] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ff6b3f-664d-4d13-b6a8-face27064f61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.778435] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf/68268135-4cf4-4732-9104-1720bbb21acf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.778755] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a6c481b-3d84-4796-95c6-2f488c2de261 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.807212] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 932.807212] env[61991]: value = "task-1130097" [ 932.807212] env[61991]: _type = "Task" [ 932.807212] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.816380] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.900758] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130092, 'name': CreateVM_Task, 'duration_secs': 1.581621} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.903210] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.903968] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.904203] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.904559] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.905171] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-605f91cb-2770-4016-932b-afbdae72a3d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.910781] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 932.910781] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529f7227-e51f-9997-189d-700ee72f1fed" [ 932.910781] env[61991]: _type = "Task" [ 932.910781] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.920502] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529f7227-e51f-9997-189d-700ee72f1fed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.970939] env[61991]: DEBUG nova.network.neutron [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updated VIF entry in instance network info cache for port e0abb70f-8b4c-4450-a784-ec0dd2381264. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 932.971410] env[61991]: DEBUG nova.network.neutron [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.039648] env[61991]: DEBUG oslo_concurrency.lockutils [req-17b8b13e-7ff9-47ff-99bb-25d2e6cebd11 req-141f9535-709b-433a-9402-6261ef654083 service nova] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.126564] env[61991]: DEBUG nova.network.neutron [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Successfully updated port: b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.318398] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.421425] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529f7227-e51f-9997-189d-700ee72f1fed, 'name': SearchDatastore_Task, 'duration_secs': 0.010951} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.421736] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.421990] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.422265] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.422482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.422678] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.422887] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-352b8883-3f68-4998-87e0-c8c1ff57b353 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.438577] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.438777] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 933.439521] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b1cb77b-02ee-4e28-9265-f97b91b1386a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.445449] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 933.445449] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dccafa-5df4-3b2d-9942-61ad4cb36b61" [ 933.445449] env[61991]: _type = "Task" [ 933.445449] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.454546] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dccafa-5df4-3b2d-9942-61ad4cb36b61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.473959] env[61991]: DEBUG oslo_concurrency.lockutils [req-4474392c-08cc-4752-be16-1f6a180d5e08 req-f2a662c5-b003-494a-8981-fe9a90672e4c service nova] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.630258] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.630484] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.631762] env[61991]: DEBUG nova.network.neutron [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.707031] env[61991]: DEBUG nova.objects.instance [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'flavor' on Instance uuid b3a2ea40-5c4c-4e7d-95b1-38a18c429cec {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.819386] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130097, 'name': ReconfigVM_Task, 'duration_secs': 0.77089} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.819699] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf/68268135-4cf4-4732-9104-1720bbb21acf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.820361] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae580d11-71f1-4ff1-973a-67eb21bbcc09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.828616] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 933.828616] env[61991]: value = "task-1130098" [ 933.828616] env[61991]: _type = "Task" [ 933.828616] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.837722] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130098, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.956769] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dccafa-5df4-3b2d-9942-61ad4cb36b61, 'name': SearchDatastore_Task, 'duration_secs': 0.017821} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.957624] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c29b6e68-0a4a-4e0e-915b-ed0b3705335a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.966106] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 933.966106] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dc2d06-8c9b-abe8-6cb8-ba4fa8e9c1ec" [ 933.966106] env[61991]: _type = "Task" [ 933.966106] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.978611] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dc2d06-8c9b-abe8-6cb8-ba4fa8e9c1ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.178977] env[61991]: WARNING nova.network.neutron [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] e6f9d471-4f42-4781-93c3-f4a9ff84317b already exists in list: networks containing: ['e6f9d471-4f42-4781-93c3-f4a9ff84317b']. ignoring it [ 934.212173] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1f5731d6-5c1e-424f-858b-8824ceb3affd tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.795s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.343338] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130098, 'name': Rename_Task, 'duration_secs': 0.193893} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.343338] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.343338] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-982c195a-ab0b-447d-9390-66c16daa52e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.352207] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 934.352207] env[61991]: value = "task-1130099" [ 934.352207] env[61991]: _type = "Task" [ 934.352207] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.361689] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.484175] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52dc2d06-8c9b-abe8-6cb8-ba4fa8e9c1ec, 'name': SearchDatastore_Task, 'duration_secs': 0.01312} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.484175] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.484175] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 934.484175] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2fe7fc7-2bd0-4499-aea4-3429baf63099 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.492360] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 934.492360] env[61991]: value = "task-1130100" [ 934.492360] env[61991]: _type = "Task" [ 934.492360] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.504965] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.574582] env[61991]: DEBUG nova.network.neutron [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "address": "fa:16:3e:2d:ad:50", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb603ad92-a2", "ovs_interfaceid": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.865110] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130099, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.006103] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130100, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.077293] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.078032] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.078211] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.079106] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04e3312-b296-439a-8d3d-329057087c1a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.097318] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.097557] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.097675] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.097870] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.098032] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.098239] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.098525] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.098759] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.099018] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.099281] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.099539] env[61991]: DEBUG nova.virt.hardware [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.106528] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Reconfiguring VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 935.107015] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6f06fd4-d0c6-4232-8a98-904533627aa0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.127967] env[61991]: DEBUG oslo_vmware.api [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 935.127967] env[61991]: value = "task-1130101" [ 935.127967] env[61991]: _type = "Task" [ 935.127967] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.137171] env[61991]: DEBUG oslo_vmware.api [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130101, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.238927] env[61991]: DEBUG nova.compute.manager [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-vif-plugged-b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.239310] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.239310] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.239432] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.239724] env[61991]: DEBUG nova.compute.manager [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] No waiting events found dispatching network-vif-plugged-b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.239807] env[61991]: WARNING nova.compute.manager [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received unexpected event network-vif-plugged-b603ad92-a23f-4da6-b5dc-60d547ba32ae for instance with vm_state active and task_state None. [ 935.239996] env[61991]: DEBUG nova.compute.manager [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-changed-b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.240203] env[61991]: DEBUG nova.compute.manager [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing instance network info cache due to event network-changed-b603ad92-a23f-4da6-b5dc-60d547ba32ae. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.240353] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.240512] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.240663] env[61991]: DEBUG nova.network.neutron [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Refreshing network info cache for port b603ad92-a23f-4da6-b5dc-60d547ba32ae {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.365362] env[61991]: DEBUG oslo_vmware.api [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130099, 'name': PowerOnVM_Task, 'duration_secs': 0.803312} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.365447] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.365681] env[61991]: INFO nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Took 9.33 seconds to spawn the instance on the hypervisor. [ 935.365930] env[61991]: DEBUG nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.366935] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0895c52-c618-4aa1-8287-92f87e03f291 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.391941] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "7ad63ade-cff7-4a32-b567-783db726de55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.391941] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "7ad63ade-cff7-4a32-b567-783db726de55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.434344] env[61991]: DEBUG nova.compute.manager [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 935.508852] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616619} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.509194] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.509559] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.509945] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78ccba9d-93d0-4934-9831-53b0d8be7536 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.520429] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 935.520429] env[61991]: value = "task-1130102" [ 935.520429] env[61991]: _type = "Task" [ 935.520429] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.533218] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.640147] env[61991]: DEBUG oslo_vmware.api [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130101, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.732068] env[61991]: INFO nova.compute.manager [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Rebuilding instance [ 935.782036] env[61991]: DEBUG nova.compute.manager [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.782943] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e1cfc9-4d92-4961-8b86-6b5c7a5bb926 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.886968] env[61991]: INFO nova.compute.manager [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Took 16.81 seconds to build instance. [ 935.894705] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.952643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.953015] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.003109] env[61991]: DEBUG nova.network.neutron [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updated VIF entry in instance network info cache for port b603ad92-a23f-4da6-b5dc-60d547ba32ae. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.003658] env[61991]: DEBUG nova.network.neutron [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "address": "fa:16:3e:2d:ad:50", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb603ad92-a2", "ovs_interfaceid": "b603ad92-a23f-4da6-b5dc-60d547ba32ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.030374] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.280079} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.030650] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.031443] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44d10c0-fbf2-4790-b916-93bd630f183c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.054244] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.054847] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1074420-e86f-49de-8ab4-d5edc870d628 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.077097] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 936.077097] env[61991]: value = "task-1130103" [ 936.077097] env[61991]: _type = "Task" [ 936.077097] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.143382] env[61991]: DEBUG oslo_vmware.api [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130101, 'name': ReconfigVM_Task, 'duration_secs': 0.676301} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.144072] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.144320] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Reconfigured VM to attach interface {{(pid=61991) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 936.298583] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.298925] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4bcd744-d7b9-4ce5-b19a-32e51c0cd83d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.308964] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 936.308964] env[61991]: value = "task-1130104" [ 936.308964] env[61991]: _type = "Task" [ 936.308964] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.318148] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130104, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.389077] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6d74413-899f-4fdd-a981-0791b1b7f804 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.318s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.421737] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.460198] env[61991]: INFO nova.compute.claims [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.508404] env[61991]: DEBUG oslo_concurrency.lockutils [req-ba29a73f-5641-42d0-b316-a7bb9217d199 req-8da0f2b7-debd-413d-9e85-4d4f2fbbe56b service nova] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.589218] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130103, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.649350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5f06acfc-0029-4881-885d-ad27b94a3026 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-902f158c-fcc4-40a2-abbd-491bdfd4b77a-b603ad92-a23f-4da6-b5dc-60d547ba32ae" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.310s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.821208] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130104, 'name': PowerOffVM_Task, 'duration_secs': 0.24574} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.821550] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.889338] env[61991]: INFO nova.compute.manager [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Detaching volume e1d09948-3670-4f76-ad45-f10a94df1f32 [ 936.930153] env[61991]: INFO nova.virt.block_device [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Attempting to driver detach volume e1d09948-3670-4f76-ad45-f10a94df1f32 from mountpoint /dev/sdb [ 936.930383] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 936.930653] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 936.931605] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a317afa-4297-41b8-a476-d7b3be819e7e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.957619] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5961235-66f3-4e69-b569-47435a5b8b8b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.966823] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0620fc7c-9622-4750-be96-753a0935ab4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.971408] env[61991]: INFO nova.compute.resource_tracker [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating resource usage from migration f5a2581b-5e86-49df-9a62-2b362ef342c1 [ 936.994309] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9615d9ed-101f-44cb-b7a3-39c25920aa90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.014901] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] The volume has not been displaced from its original location: [datastore1] volume-e1d09948-3670-4f76-ad45-f10a94df1f32/volume-e1d09948-3670-4f76-ad45-f10a94df1f32.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 937.020435] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfiguring VM instance instance-00000054 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 937.023525] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7cce8e1-012e-4330-bae8-7a5d40a1311c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.045263] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 937.045263] env[61991]: value = "task-1130105" [ 937.045263] env[61991]: _type = "Task" [ 937.045263] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.058040] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130105, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.090072] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130103, 'name': ReconfigVM_Task, 'duration_secs': 0.537991} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.090311] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.093529] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b8f7b76-c6da-4555-9de5-93dea9e695b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.102947] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 937.102947] env[61991]: value = "task-1130106" [ 937.102947] env[61991]: _type = "Task" [ 937.102947] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.115346] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130106, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.321475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb12ef9-5d53-4447-9c24-c09c9e486541 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.330327] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb5f91d-9aa9-4278-ad16-94460e85d4cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.364545] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc3b45e-4392-4f26-9910-3d2d622fe563 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.373620] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e18123-a4d8-472b-8f36-c04e7f68febd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.388066] env[61991]: DEBUG nova.compute.provider_tree [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.556905] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130105, 'name': ReconfigVM_Task, 'duration_secs': 0.274525} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.556905] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfigured VM instance instance-00000054 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 937.561730] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7806d54b-eb5d-416e-bba0-3687c733aace {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.583291] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 937.583291] env[61991]: value = "task-1130107" [ 937.583291] env[61991]: _type = "Task" [ 937.583291] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.593121] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130107, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.612908] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130106, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.892784] env[61991]: DEBUG nova.scheduler.client.report [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.090537] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52289603-565b-2917-c64e-131de310bc9f/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 938.094411] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3f8a97-6391-4e4f-941e-6c29d1e3c9f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.109497] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52289603-565b-2917-c64e-131de310bc9f/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 938.109497] env[61991]: ERROR oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52289603-565b-2917-c64e-131de310bc9f/disk-0.vmdk due to incomplete transfer. [ 938.109497] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130107, 'name': ReconfigVM_Task, 'duration_secs': 0.168511} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.112765] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-23f27234-b0dc-460a-902f-be7d87383fa9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.114347] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 938.123536] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130106, 'name': Rename_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.125025] env[61991]: DEBUG oslo_vmware.rw_handles [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52289603-565b-2917-c64e-131de310bc9f/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 938.126302] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Uploaded image 5cf3221b-0c05-4b11-a46c-192b39b794ad to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 938.129514] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 938.129618] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-76ff1fc6-3514-468a-85ee-664325040fef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.141920] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 938.141920] env[61991]: value = "task-1130108" [ 938.141920] env[61991]: _type = "Task" [ 938.141920] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.152813] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130108, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.290711] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "interface-902f158c-fcc4-40a2-abbd-491bdfd4b77a-b603ad92-a23f-4da6-b5dc-60d547ba32ae" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.291166] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-902f158c-fcc4-40a2-abbd-491bdfd4b77a-b603ad92-a23f-4da6-b5dc-60d547ba32ae" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.359493] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "077c5def-84cc-48f9-be90-410100666773" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.359798] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "077c5def-84cc-48f9-be90-410100666773" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.360010] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "077c5def-84cc-48f9-be90-410100666773-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.360231] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "077c5def-84cc-48f9-be90-410100666773-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.360385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "077c5def-84cc-48f9-be90-410100666773-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.362976] env[61991]: INFO nova.compute.manager [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Terminating instance [ 938.370190] env[61991]: DEBUG nova.compute.manager [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.371761] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.374021] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3425d86a-0fb8-4ce6-910c-1c6011b7d6b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.387231] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.387528] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c66dbb3a-595d-420c-89a5-e33cfe772dd4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.398364] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.445s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.398584] env[61991]: INFO nova.compute.manager [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Migrating [ 938.408701] env[61991]: DEBUG oslo_vmware.api [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 938.408701] env[61991]: value = "task-1130109" [ 938.408701] env[61991]: _type = "Task" [ 938.408701] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.409138] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.988s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.412517] env[61991]: INFO nova.compute.claims [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.434208] env[61991]: DEBUG oslo_vmware.api [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130109, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.622295] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130106, 'name': Rename_Task, 'duration_secs': 1.249365} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.622295] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.622295] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40a2334e-5442-4d3d-832c-93b84ba74a6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.632597] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 938.632597] env[61991]: value = "task-1130110" [ 938.632597] env[61991]: _type = "Task" [ 938.632597] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.648316] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130110, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.658422] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130108, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.795732] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.796048] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.797132] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3accf370-aaf3-4f7a-8ffc-d4ebcb0a128a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.824529] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d39a16-870f-47ca-a9b1-9f71542809f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.857542] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Reconfiguring VM to detach interface {{(pid=61991) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 938.857831] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fa9e8cd-a197-433c-8b5e-4131f46c0326 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.897628] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 938.897628] env[61991]: value = "task-1130111" [ 938.897628] env[61991]: _type = "Task" [ 938.897628] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.908349] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.925813] env[61991]: DEBUG oslo_vmware.api [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130109, 'name': PowerOffVM_Task, 'duration_secs': 0.509863} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.926281] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.926511] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.926825] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c470b4c3-a783-46e8-aa22-5a63292820ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.931716] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.932035] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.932283] env[61991]: DEBUG nova.network.neutron [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 938.946163] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.946491] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.946853] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.946951] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.947232] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.950311] env[61991]: INFO nova.compute.manager [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Terminating instance [ 938.952810] env[61991]: DEBUG nova.compute.manager [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.953153] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.954058] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298b00e8-eb17-43c2-a6d1-24404fea4d2e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.964804] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.965235] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11ff1365-5110-46c1-9a8a-56bde9734ad0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.977710] env[61991]: DEBUG oslo_vmware.api [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 938.977710] env[61991]: value = "task-1130113" [ 938.977710] env[61991]: _type = "Task" [ 938.977710] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.988749] env[61991]: DEBUG oslo_vmware.api [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130113, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.054735] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.055503] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.055503] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleting the datastore file [datastore1] 077c5def-84cc-48f9-be90-410100666773 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.055615] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6da5cadb-2b98-4e0a-b28c-3b72a094e675 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.064260] env[61991]: DEBUG oslo_vmware.api [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 939.064260] env[61991]: value = "task-1130114" [ 939.064260] env[61991]: _type = "Task" [ 939.064260] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.074747] env[61991]: DEBUG oslo_vmware.api [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130114, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.149526] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130110, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.164602] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130108, 'name': Destroy_Task, 'duration_secs': 0.866249} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.165040] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Destroyed the VM [ 939.165415] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 939.165976] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d099a660-23e1-4202-a0ec-227a8af1a6ea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.179202] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.179784] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a507295a-27c8-47cc-a8c5-2b26d8148571 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.184734] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 939.184734] env[61991]: value = "task-1130115" [ 939.184734] env[61991]: _type = "Task" [ 939.184734] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.194218] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 939.194218] env[61991]: value = "task-1130116" [ 939.194218] env[61991]: _type = "Task" [ 939.194218] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.204903] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130115, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.216299] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 939.216686] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 939.217015] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 939.218342] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae5dfaf-a9ce-469d-90f0-268329de73de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.254287] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b621b83c-6c3d-4a3e-aa7f-0bc95fb2f656 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.266437] env[61991]: WARNING nova.virt.vmwareapi.driver [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 939.266982] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.268365] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c161fcaa-6331-4208-bf9e-71e68ed4ed30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.279438] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.279826] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12803185-8f6f-467a-b4a7-1858b29fd989 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.312577] env[61991]: DEBUG nova.compute.manager [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 939.404187] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.404685] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.404685] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.404882] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3dcda5a7-2b9b-4f15-9a0c-dda10473e071 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.410911] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.418974] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 939.418974] env[61991]: value = "task-1130118" [ 939.418974] env[61991]: _type = "Task" [ 939.418974] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.429493] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130118, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.499930] env[61991]: DEBUG oslo_vmware.api [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130113, 'name': PowerOffVM_Task, 'duration_secs': 0.277968} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.503892] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.504456] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.509148] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de0b90f6-515d-4172-a361-1b3c7c0f00ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.579222] env[61991]: DEBUG oslo_vmware.api [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130114, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267154} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.583535] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.583884] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.584458] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.584557] env[61991]: INFO nova.compute.manager [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 077c5def-84cc-48f9-be90-410100666773] Took 1.21 seconds to destroy the instance on the hypervisor. [ 939.584883] env[61991]: DEBUG oslo.service.loopingcall [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.585367] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.585709] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.586060] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.586386] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.586674] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.588966] env[61991]: DEBUG nova.compute.manager [-] [instance: 077c5def-84cc-48f9-be90-410100666773] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.589093] env[61991]: DEBUG nova.network.neutron [-] [instance: 077c5def-84cc-48f9-be90-410100666773] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.591162] env[61991]: INFO nova.compute.manager [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Terminating instance [ 939.593332] env[61991]: DEBUG nova.compute.manager [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.593531] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.594381] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d650c0-88c0-45b9-ae54-64cfee178f82 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.609091] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.609458] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.609657] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.610116] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Deleting the datastore file [datastore2] 586bf51a-d1c4-44e8-863b-b229a3ca4f2a {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.610116] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1227a2a5-8b79-4886-9e1c-a748c9ca1aa5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.611837] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5fb02154-c4b1-4cae-a05b-c1fdc4c3c311 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.619896] env[61991]: DEBUG oslo_vmware.api [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for the task: (returnval){ [ 939.619896] env[61991]: value = "task-1130121" [ 939.619896] env[61991]: _type = "Task" [ 939.619896] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.624460] env[61991]: DEBUG oslo_vmware.api [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 939.624460] env[61991]: value = "task-1130120" [ 939.624460] env[61991]: _type = "Task" [ 939.624460] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.648615] env[61991]: DEBUG oslo_vmware.api [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.648915] env[61991]: DEBUG oslo_vmware.api [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.656350] env[61991]: DEBUG oslo_vmware.api [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130110, 'name': PowerOnVM_Task, 'duration_secs': 0.925451} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.656681] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.656910] env[61991]: INFO nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Took 11.18 seconds to spawn the instance on the hypervisor. [ 939.657123] env[61991]: DEBUG nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 939.660554] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae254cbe-2196-4720-96f5-1139624848df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.698105] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130115, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.836571] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.858453] env[61991]: DEBUG nova.network.neutron [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance_info_cache with network_info: [{"id": "2dc3450f-efd8-4200-a182-988386f68d00", "address": "fa:16:3e:de:38:21", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dc3450f-ef", "ovs_interfaceid": "2dc3450f-efd8-4200-a182-988386f68d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.883087] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fd48a8-66da-4179-b5dc-72aa5184c93f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.893117] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e715008c-e2e5-4c10-b0a0-b3b5e75ff124 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.938902] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92cbfff-60b3-4f41-b5ef-e402e441b5f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.942280] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.952087] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8f0d85-7dde-4cf5-b9b3-bcd3094ac2ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.956255] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130118, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300781} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.956590] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.956743] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.956992] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.972311] env[61991]: DEBUG nova.compute.provider_tree [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.084466] env[61991]: DEBUG nova.compute.manager [req-3bd2f34a-74b6-4b8d-9679-c3de8555c8eb req-791e2382-a351-4524-94a5-3e7a6e33df99 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Received event network-vif-deleted-2ca5206e-8353-4094-b7cd-257f04c3fdcb {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.085357] env[61991]: INFO nova.compute.manager [req-3bd2f34a-74b6-4b8d-9679-c3de8555c8eb req-791e2382-a351-4524-94a5-3e7a6e33df99 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Neutron deleted interface 2ca5206e-8353-4094-b7cd-257f04c3fdcb; detaching it from the instance and deleting it from the info cache [ 940.085357] env[61991]: DEBUG nova.network.neutron [req-3bd2f34a-74b6-4b8d-9679-c3de8555c8eb req-791e2382-a351-4524-94a5-3e7a6e33df99 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.136967] env[61991]: DEBUG oslo_vmware.api [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Task: {'id': task-1130121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277635} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.136967] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.137266] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.137448] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.137631] env[61991]: INFO nova.compute.manager [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Took 1.18 seconds to destroy the instance on the hypervisor. [ 940.137874] env[61991]: DEBUG oslo.service.loopingcall [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.138070] env[61991]: DEBUG nova.compute.manager [-] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.138235] env[61991]: DEBUG nova.network.neutron [-] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.142950] env[61991]: DEBUG oslo_vmware.api [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130120, 'name': PowerOffVM_Task, 'duration_secs': 0.247046} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.143508] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.144364] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 940.144364] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-101a3f12-ee1e-4f2e-9f40-9ea351ba8388 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.195031] env[61991]: INFO nova.compute.manager [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Took 19.94 seconds to build instance. [ 940.199741] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130115, 'name': RemoveSnapshot_Task, 'duration_secs': 0.684044} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.200495] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 940.200788] env[61991]: DEBUG nova.compute.manager [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.201806] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55146430-b716-47b8-a976-520c7fad977d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.224345] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 940.224885] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 940.225235] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Deleting the datastore file [datastore1] d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.225581] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3de0dd18-556e-4922-95a2-e13e954f5f49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.235946] env[61991]: DEBUG oslo_vmware.api [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 940.235946] env[61991]: value = "task-1130123" [ 940.235946] env[61991]: _type = "Task" [ 940.235946] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.246026] env[61991]: DEBUG oslo_vmware.api [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.364708] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.414183] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.464437] env[61991]: INFO nova.virt.block_device [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Booting with volume e1d09948-3670-4f76-ad45-f10a94df1f32 at /dev/sdb [ 940.475405] env[61991]: DEBUG nova.scheduler.client.report [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.500907] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c20581f-97b6-4091-b7f7-f948f2151512 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.511867] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efce0e73-de71-46fe-8d3d-c1d2c8ae303c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.550083] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f958bab-004a-4afd-aab7-67fdebfaa1aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.561574] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ac133e-1739-4a61-a377-79446c219e95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.573903] env[61991]: DEBUG nova.network.neutron [-] [instance: 077c5def-84cc-48f9-be90-410100666773] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.588212] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd3f1c5c-7a93-4ce7-a75f-7b7e3e5ffbfa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.606214] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e934a05e-aeb7-4ac9-b965-bc2ea84ab3c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.613646] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cb42b2-d49b-4781-a5db-c9bd49145dc6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.631012] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef904a09-f937-44f3-a851-3c8e423870a2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.659983] env[61991]: DEBUG nova.compute.manager [req-3bd2f34a-74b6-4b8d-9679-c3de8555c8eb req-791e2382-a351-4524-94a5-3e7a6e33df99 service nova] [instance: 077c5def-84cc-48f9-be90-410100666773] Detach interface failed, port_id=2ca5206e-8353-4094-b7cd-257f04c3fdcb, reason: Instance 077c5def-84cc-48f9-be90-410100666773 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 940.665694] env[61991]: DEBUG nova.virt.block_device [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating existing volume attachment record: cadb458b-d45b-49d3-a36e-623a9bdb1c4d {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 940.676328] env[61991]: DEBUG nova.compute.manager [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Received event network-changed-4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.677268] env[61991]: DEBUG nova.compute.manager [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Refreshing instance network info cache due to event network-changed-4e506178-06a6-49c6-a6b3-fa24ebafc988. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.677268] env[61991]: DEBUG oslo_concurrency.lockutils [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.677655] env[61991]: DEBUG oslo_concurrency.lockutils [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.678806] env[61991]: DEBUG nova.network.neutron [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Refreshing network info cache for port 4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 940.701568] env[61991]: DEBUG oslo_concurrency.lockutils [None req-cc3be1c7-aa56-42de-865b-004438332de4 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.479s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.718454] env[61991]: INFO nova.compute.manager [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Shelve offloading [ 940.722057] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.722057] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95c52536-d11b-49ac-bbbe-a6452f0bf672 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.730693] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 940.730693] env[61991]: value = "task-1130124" [ 940.730693] env[61991]: _type = "Task" [ 940.730693] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.739697] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130124, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.748889] env[61991]: DEBUG oslo_vmware.api [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186995} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.749160] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.749366] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.749548] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.749687] env[61991]: INFO nova.compute.manager [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Took 1.16 seconds to destroy the instance on the hypervisor. [ 940.749922] env[61991]: DEBUG oslo.service.loopingcall [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.750116] env[61991]: DEBUG nova.compute.manager [-] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.750209] env[61991]: DEBUG nova.network.neutron [-] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.878869] env[61991]: DEBUG nova.network.neutron [-] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.915550] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.982174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.982797] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.988039] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.152s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.077091] env[61991]: INFO nova.compute.manager [-] [instance: 077c5def-84cc-48f9-be90-410100666773] Took 1.49 seconds to deallocate network for instance. [ 941.242560] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 941.242560] env[61991]: DEBUG nova.compute.manager [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.243310] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae7f8a7-9d00-402a-8c78-719387a85ad9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.250651] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.251069] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.251069] env[61991]: DEBUG nova.network.neutron [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.384129] env[61991]: INFO nova.compute.manager [-] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Took 1.24 seconds to deallocate network for instance. [ 941.413162] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.461364] env[61991]: DEBUG nova.network.neutron [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updated VIF entry in instance network info cache for port 4e506178-06a6-49c6-a6b3-fa24ebafc988. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 941.461741] env[61991]: DEBUG nova.network.neutron [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.491405] env[61991]: DEBUG nova.compute.utils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 941.492908] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 941.494344] env[61991]: DEBUG nova.network.neutron [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 941.499041] env[61991]: INFO nova.compute.claims [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.545384] env[61991]: DEBUG nova.policy [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceec39441db345a1ac9cc3f9d6f6ee18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '870d2c2c0e554180b190b88bdab5fc2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 941.549343] env[61991]: DEBUG nova.network.neutron [-] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.583297] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.834047] env[61991]: DEBUG nova.network.neutron [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Successfully created port: f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.879647] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d5c52d-3863-4167-a56f-024c1ff927eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.900757] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.900757] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.918947] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.964784] env[61991]: DEBUG oslo_concurrency.lockutils [req-6bb25c07-39e1-4064-a0f3-7627e79c4c01 req-4063a84f-d50b-4434-a23a-98568ce2f457 service nova] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.997312] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 942.004442] env[61991]: INFO nova.compute.resource_tracker [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating resource usage from migration ac7f2f11-e6d1-4f68-81ba-58a9c7b870d1 [ 942.035867] env[61991]: DEBUG nova.network.neutron [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.052321] env[61991]: INFO nova.compute.manager [-] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Took 1.30 seconds to deallocate network for instance. [ 942.169949] env[61991]: DEBUG nova.compute.manager [req-620809d2-af16-47ae-a16f-7781f088665b req-f6ec6240-130f-4997-bc31-1e68803cd28c service nova] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Received event network-vif-deleted-7cc61fcf-f52c-468a-8d9d-91635e8e25e9 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.170268] env[61991]: DEBUG nova.compute.manager [req-620809d2-af16-47ae-a16f-7781f088665b req-f6ec6240-130f-4997-bc31-1e68803cd28c service nova] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Received event network-vif-deleted-7049a455-7d87-4402-a422-05ce15c9264b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.325874] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6993fef-a237-4be7-b59d-ec5aab0ef8ef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.335352] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d058ac-9bb1-48b1-b10a-b2486f8adac5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.367061] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fe1c32-2056-4609-a6c5-ced6d9f727ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.375044] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb048cde-6a84-4cbd-9213-84c6c621b75a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.388579] env[61991]: DEBUG nova.compute.provider_tree [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.409423] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.414068] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc7141b5-2191-40f3-8091-ca0c984e17dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.422792] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.424196] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 942.424196] env[61991]: value = "task-1130125" [ 942.424196] env[61991]: _type = "Task" [ 942.424196] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.432709] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.540537] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.565683] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.793736] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.793997] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.797276] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.797495] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.797651] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.797805] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.798042] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.798213] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.798382] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.799871] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.800106] env[61991]: DEBUG nova.virt.hardware [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.801592] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c552e5-aede-4cb5-bbcd-f89b3ac811fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.805736] env[61991]: DEBUG nova.compute.manager [req-0b1d356d-4c9d-4a1b-a6c1-1e9590fa9d6a req-086d26ec-17b2-4baf-a553-14b8b6b4033d service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-vif-unplugged-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.805967] env[61991]: DEBUG oslo_concurrency.lockutils [req-0b1d356d-4c9d-4a1b-a6c1-1e9590fa9d6a req-086d26ec-17b2-4baf-a553-14b8b6b4033d service nova] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.806205] env[61991]: DEBUG oslo_concurrency.lockutils [req-0b1d356d-4c9d-4a1b-a6c1-1e9590fa9d6a req-086d26ec-17b2-4baf-a553-14b8b6b4033d service nova] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.806377] env[61991]: DEBUG oslo_concurrency.lockutils [req-0b1d356d-4c9d-4a1b-a6c1-1e9590fa9d6a req-086d26ec-17b2-4baf-a553-14b8b6b4033d service nova] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.806545] env[61991]: DEBUG nova.compute.manager [req-0b1d356d-4c9d-4a1b-a6c1-1e9590fa9d6a req-086d26ec-17b2-4baf-a553-14b8b6b4033d service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] No waiting events found dispatching network-vif-unplugged-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.806711] env[61991]: WARNING nova.compute.manager [req-0b1d356d-4c9d-4a1b-a6c1-1e9590fa9d6a req-086d26ec-17b2-4baf-a553-14b8b6b4033d service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received unexpected event network-vif-unplugged-c7263e46-a164-4813-be3e-faae7fc31ff2 for instance with vm_state shelved and task_state shelving_offloading. [ 942.813755] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f5f73c-a605-4222-9932-b030a458f3ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.829501] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:8a:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.837083] env[61991]: DEBUG oslo.service.loopingcall [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.837432] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.837721] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-168f6411-4f52-4c03-9361-e0ceae6d77a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.853169] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.854283] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8cd48f-b997-46d6-8559-8681d8a6c63a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.862558] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.863854] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12c83079-026b-4d47-86f8-d34e341e4930 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.865261] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.865261] env[61991]: value = "task-1130126" [ 942.865261] env[61991]: _type = "Task" [ 942.865261] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.873062] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130126, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.892190] env[61991]: DEBUG nova.scheduler.client.report [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.918385] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.929735] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.929975] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.930154] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleting the datastore file [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.930976] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cd7596c-7eb2-4b69-a975-e275e8ba4147 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.936467] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130125, 'name': PowerOffVM_Task, 'duration_secs': 0.21364} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.937685] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.937874] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 942.941260] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 942.941260] env[61991]: value = "task-1130128" [ 942.941260] env[61991]: _type = "Task" [ 942.941260] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.949810] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.008123] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 943.029493] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.029767] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.029933] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.030149] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.030299] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.030442] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.030650] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.030812] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.031025] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.031459] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.032109] env[61991]: DEBUG nova.virt.hardware [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.032969] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3048f3ea-9d93-4cf3-843b-6204b0922389 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.042435] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ce4af0-28c6-4c9f-b32d-1ed6a5b975bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.376233] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130126, 'name': CreateVM_Task, 'duration_secs': 0.461938} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.376233] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.376797] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.377047] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.377398] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.377683] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5451ef83-c24d-417e-b908-d4764c02d88e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.384091] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 943.384091] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5216b7c1-e1fb-c58a-d3ee-fb03b70b09cb" [ 943.384091] env[61991]: _type = "Task" [ 943.384091] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.392381] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5216b7c1-e1fb-c58a-d3ee-fb03b70b09cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.397407] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.409s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.401021] env[61991]: INFO nova.compute.manager [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Migrating [ 943.407505] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.824s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.407779] env[61991]: DEBUG nova.objects.instance [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lazy-loading 'resources' on Instance uuid 077c5def-84cc-48f9-be90-410100666773 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.431167] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.440089] env[61991]: DEBUG nova.compute.manager [req-93fe31f5-9785-4793-aeb6-6ac272974882 req-7ba528bc-9e11-4429-bbfc-4edf6824d0fa service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Received event network-vif-plugged-f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.440366] env[61991]: DEBUG oslo_concurrency.lockutils [req-93fe31f5-9785-4793-aeb6-6ac272974882 req-7ba528bc-9e11-4429-bbfc-4edf6824d0fa service nova] Acquiring lock "7ad63ade-cff7-4a32-b567-783db726de55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.440522] env[61991]: DEBUG oslo_concurrency.lockutils [req-93fe31f5-9785-4793-aeb6-6ac272974882 req-7ba528bc-9e11-4429-bbfc-4edf6824d0fa service nova] Lock "7ad63ade-cff7-4a32-b567-783db726de55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.440688] env[61991]: DEBUG oslo_concurrency.lockutils [req-93fe31f5-9785-4793-aeb6-6ac272974882 req-7ba528bc-9e11-4429-bbfc-4edf6824d0fa service nova] Lock "7ad63ade-cff7-4a32-b567-783db726de55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.440852] env[61991]: DEBUG nova.compute.manager [req-93fe31f5-9785-4793-aeb6-6ac272974882 req-7ba528bc-9e11-4429-bbfc-4edf6824d0fa service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] No waiting events found dispatching network-vif-plugged-f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.441032] env[61991]: WARNING nova.compute.manager [req-93fe31f5-9785-4793-aeb6-6ac272974882 req-7ba528bc-9e11-4429-bbfc-4edf6824d0fa service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Received unexpected event network-vif-plugged-f039d7a6-60dd-49c4-95bc-ba71796b1aba for instance with vm_state building and task_state spawning. [ 943.443675] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.443924] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.444102] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.444295] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.444459] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.448024] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.448024] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.448024] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.448024] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.448024] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.448024] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.450984] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53d20c98-0f34-4a3f-abf6-30d5ddc7c634 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.472263] env[61991]: DEBUG oslo_vmware.api [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153669} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.474025] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.474442] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.474629] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.477796] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 943.477796] env[61991]: value = "task-1130129" [ 943.477796] env[61991]: _type = "Task" [ 943.477796] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.489836] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130129, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.507712] env[61991]: INFO nova.scheduler.client.report [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted allocations for instance 56188232-bed0-4a4b-a4bc-01edbb85cbe4 [ 943.560019] env[61991]: DEBUG nova.network.neutron [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Successfully updated port: f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.893866] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5216b7c1-e1fb-c58a-d3ee-fb03b70b09cb, 'name': SearchDatastore_Task, 'duration_secs': 0.011998} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.894220] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.894479] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.894726] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.894878] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.895083] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.895362] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a10a3add-76cf-4788-a579-2658f3c1cac7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.905365] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.905549] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.906316] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f9f3c61-84fb-4c54-a118-b3033df76cad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.912710] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 943.912710] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a0c549-5d3a-73d7-4c22-12efb2963b86" [ 943.912710] env[61991]: _type = "Task" [ 943.912710] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.915873] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.916074] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.916287] env[61991]: DEBUG nova.network.neutron [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.927775] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a0c549-5d3a-73d7-4c22-12efb2963b86, 'name': SearchDatastore_Task, 'duration_secs': 0.009018} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.933995] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-706d90e6-6563-4947-badf-478c45abc1c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.943287] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.950904] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 943.950904] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5211798e-3d8d-c2e2-0fb2-ddcb1d1ab0f0" [ 943.950904] env[61991]: _type = "Task" [ 943.950904] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.961835] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5211798e-3d8d-c2e2-0fb2-ddcb1d1ab0f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.989853] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130129, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.012894] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.062350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-7ad63ade-cff7-4a32-b567-783db726de55" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.062495] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-7ad63ade-cff7-4a32-b567-783db726de55" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.062644] env[61991]: DEBUG nova.network.neutron [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.161421] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8b1794-ca61-41e4-8637-5b017e3d641d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.170104] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7e5212-8667-4a50-9839-ae3c9539a44e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.202698] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78631be-6ba1-4d40-9f26-84a3f4df4a30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.206188] env[61991]: DEBUG nova.compute.manager [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Received event network-changed-f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.206390] env[61991]: DEBUG nova.compute.manager [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Refreshing instance network info cache due to event network-changed-f039d7a6-60dd-49c4-95bc-ba71796b1aba. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.206588] env[61991]: DEBUG oslo_concurrency.lockutils [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] Acquiring lock "refresh_cache-7ad63ade-cff7-4a32-b567-783db726de55" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.213220] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6468420a-11aa-4cee-8664-1a110e222878 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.235509] env[61991]: DEBUG nova.compute.provider_tree [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.430597] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.464558] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5211798e-3d8d-c2e2-0fb2-ddcb1d1ab0f0, 'name': SearchDatastore_Task, 'duration_secs': 0.010462} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.464880] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.465188] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.465462] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a6a3b0b-1079-4f90-ba26-85c413a8b449 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.474170] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 944.474170] env[61991]: value = "task-1130130" [ 944.474170] env[61991]: _type = "Task" [ 944.474170] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.485235] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.490795] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130129, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.617353] env[61991]: DEBUG nova.network.neutron [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 944.716275] env[61991]: DEBUG nova.network.neutron [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance_info_cache with network_info: [{"id": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "address": "fa:16:3e:d7:45:df", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa511534a-b3", "ovs_interfaceid": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.739350] env[61991]: DEBUG nova.scheduler.client.report [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.023881] env[61991]: DEBUG oslo_vmware.api [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130111, 'name': ReconfigVM_Task, 'duration_secs': 5.785673} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.023881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.023881] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Reconfigured VM to detach interface {{(pid=61991) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 945.023881] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507214} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.023881] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.023881] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.023881] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130129, 'name': ReconfigVM_Task, 'duration_secs': 1.381156} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.023881] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76aa6d1c-571a-4111-8556-1c6584f11f61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.023881] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 945.023881] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 945.023881] env[61991]: value = "task-1130131" [ 945.023881] env[61991]: _type = "Task" [ 945.023881] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.025512] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130131, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.105195] env[61991]: DEBUG nova.compute.manager [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-changed-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.105422] env[61991]: DEBUG nova.compute.manager [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Refreshing instance network info cache due to event network-changed-c7263e46-a164-4813-be3e-faae7fc31ff2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 945.105646] env[61991]: DEBUG oslo_concurrency.lockutils [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.105797] env[61991]: DEBUG oslo_concurrency.lockutils [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.105998] env[61991]: DEBUG nova.network.neutron [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Refreshing network info cache for port c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.142956] env[61991]: DEBUG nova.network.neutron [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Updating instance_info_cache with network_info: [{"id": "f039d7a6-60dd-49c4-95bc-ba71796b1aba", "address": "fa:16:3e:6b:f3:a5", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf039d7a6-60", "ovs_interfaceid": "f039d7a6-60dd-49c4-95bc-ba71796b1aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.219281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.245355] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.248059] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.348s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.248059] env[61991]: DEBUG nova.objects.instance [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lazy-loading 'resources' on Instance uuid 586bf51a-d1c4-44e8-863b-b229a3ca4f2a {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.262436] env[61991]: INFO nova.scheduler.client.report [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleted allocations for instance 077c5def-84cc-48f9-be90-410100666773 [ 945.507907] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.508180] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.508345] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.508529] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.508677] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.508828] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.509043] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.509216] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.509386] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.509551] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.509726] env[61991]: DEBUG nova.virt.hardware [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.514924] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 945.515653] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d5cd8cf-3722-4bd5-89aa-5ca40e0ab17a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.537719] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130131, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064797} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.539063] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.539412] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 945.539412] env[61991]: value = "task-1130132" [ 945.539412] env[61991]: _type = "Task" [ 945.539412] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.540135] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a4b4e1-c017-4a9e-b661-7de342c62c0d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.568574] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.571617] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e61ba6df-9aad-4910-baac-85c539156968 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.585923] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130132, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.592054] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 945.592054] env[61991]: value = "task-1130133" [ 945.592054] env[61991]: _type = "Task" [ 945.592054] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.600660] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.645689] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-7ad63ade-cff7-4a32-b567-783db726de55" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.646081] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance network_info: |[{"id": "f039d7a6-60dd-49c4-95bc-ba71796b1aba", "address": "fa:16:3e:6b:f3:a5", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf039d7a6-60", "ovs_interfaceid": "f039d7a6-60dd-49c4-95bc-ba71796b1aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.646400] env[61991]: DEBUG oslo_concurrency.lockutils [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] Acquired lock "refresh_cache-7ad63ade-cff7-4a32-b567-783db726de55" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.646585] env[61991]: DEBUG nova.network.neutron [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Refreshing network info cache for port f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.647753] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:f3:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f039d7a6-60dd-49c4-95bc-ba71796b1aba', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.655320] env[61991]: DEBUG oslo.service.loopingcall [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.656456] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.656764] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-248f7041-3c2e-4665-81fb-65dabf30f1d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.680661] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.680661] env[61991]: value = "task-1130134" [ 945.680661] env[61991]: _type = "Task" [ 945.680661] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.691989] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130134, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.775114] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc990316-bdd3-4b6a-b5c5-b9025746e43d tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "077c5def-84cc-48f9-be90-410100666773" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.415s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.835048] env[61991]: DEBUG nova.network.neutron [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updated VIF entry in instance network info cache for port c7263e46-a164-4813-be3e-faae7fc31ff2. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 945.835435] env[61991]: DEBUG nova.network.neutron [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": null, "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc7263e46-a1", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.033869] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3213e7-e630-4d96-a93e-4fd87ebf4fac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.041885] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c9d2d8-0038-42f4-a5e7-c413c2e22330 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.053149] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130132, 'name': ReconfigVM_Task, 'duration_secs': 0.193928} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.077116] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 946.078151] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ab3486-188f-4e9a-ab3e-c0d8e5686c12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.081185] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060f4fc3-2d63-48f8-b1b7-dd59bb7ccbdb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.098844] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d11532-446c-46e3-b6f6-97454a56a6cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.109895] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514/3b647920-4a69-4d1d-adb1-6fbf716e5514.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.112873] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9756909-d296-43c8-b033-d5210aaa6a7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.144119] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.151967] env[61991]: DEBUG nova.compute.provider_tree [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.154469] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 946.154469] env[61991]: value = "task-1130135" [ 946.154469] env[61991]: _type = "Task" [ 946.154469] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.167145] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.192499] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130134, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.201034] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.341340] env[61991]: DEBUG oslo_concurrency.lockutils [req-f50e4b0c-e607-4e63-9757-69f4045f373e req-2e2c3c32-d71f-4058-9efb-b23235bb4044 service nova] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.382583] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.382791] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquired lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.382972] env[61991]: DEBUG nova.network.neutron [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 946.384558] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.386444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.386444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.386444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.386444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.388226] env[61991]: INFO nova.compute.manager [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Terminating instance [ 946.392115] env[61991]: DEBUG nova.compute.manager [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.392368] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.393858] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa81a9f-f58d-49a6-a963-0dd0d1a0cab1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.406318] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 946.406602] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0b4df80-4959-4037-8808-3a9c33e04714 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.417265] env[61991]: DEBUG oslo_vmware.api [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 946.417265] env[61991]: value = "task-1130136" [ 946.417265] env[61991]: _type = "Task" [ 946.417265] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.426647] env[61991]: DEBUG oslo_vmware.api [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.484528] env[61991]: DEBUG nova.network.neutron [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Updated VIF entry in instance network info cache for port f039d7a6-60dd-49c4-95bc-ba71796b1aba. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.484911] env[61991]: DEBUG nova.network.neutron [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Updating instance_info_cache with network_info: [{"id": "f039d7a6-60dd-49c4-95bc-ba71796b1aba", "address": "fa:16:3e:6b:f3:a5", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf039d7a6-60", "ovs_interfaceid": "f039d7a6-60dd-49c4-95bc-ba71796b1aba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.613279] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130133, 'name': ReconfigVM_Task, 'duration_secs': 0.849666} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.613579] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfigured VM instance instance-00000054 to attach disk [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec/b3a2ea40-5c4c-4e7d-95b1-38a18c429cec.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.615169] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'device_name': '/dev/sda', 'boot_index': 0, 'device_type': 'disk', 'guest_format': None, 'disk_bus': None, 'encryption_secret_uuid': None, 'encrypted': False, 'encryption_options': None, 'encryption_format': None, 'image_id': '254d700f-2f5a-49a3-8762-cec07162124a'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'guest_format': None, 'mount_device': '/dev/sdb', 'disk_bus': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'}, 'attachment_id': 'cadb458b-d45b-49d3-a36e-623a9bdb1c4d', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61991) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 946.615438] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 946.615675] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 946.616634] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861d74f0-245b-423c-8a5a-0dbd992148b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.633242] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379b4d14-86c4-4fa2-8256-d716006b0c91 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.657412] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] volume-e1d09948-3670-4f76-ad45-f10a94df1f32/volume-e1d09948-3670-4f76-ad45-f10a94df1f32.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.658424] env[61991]: DEBUG nova.scheduler.client.report [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.661674] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f70ccb48-a6ab-45f5-a9ec-1d4c1f7b55c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.676103] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.428s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.683113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.117s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.683352] env[61991]: DEBUG nova.objects.instance [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lazy-loading 'resources' on Instance uuid d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.692701] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 946.692701] env[61991]: value = "task-1130137" [ 946.692701] env[61991]: _type = "Task" [ 946.692701] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.696455] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130134, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.702394] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130135, 'name': ReconfigVM_Task, 'duration_secs': 0.319155} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.703345] env[61991]: INFO nova.scheduler.client.report [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Deleted allocations for instance 586bf51a-d1c4-44e8-863b-b229a3ca4f2a [ 946.704596] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514/3b647920-4a69-4d1d-adb1-6fbf716e5514.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.704891] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 946.718291] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.735136] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7fe82c-d00c-4387-beb2-782e3b6eaa68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.755428] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 946.877656] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.877996] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.879540] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.879540] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.879780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.883926] env[61991]: INFO nova.compute.manager [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Terminating instance [ 946.885804] env[61991]: DEBUG nova.compute.manager [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.886086] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.887431] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f224b99-59ce-4278-899c-c033d82d1c8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.898224] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 946.898479] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aaf0cf23-5245-414d-9c7f-29c760278269 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.906209] env[61991]: DEBUG oslo_vmware.api [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 946.906209] env[61991]: value = "task-1130138" [ 946.906209] env[61991]: _type = "Task" [ 946.906209] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.914657] env[61991]: DEBUG oslo_vmware.api [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.930767] env[61991]: DEBUG oslo_vmware.api [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130136, 'name': PowerOffVM_Task, 'duration_secs': 0.376196} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.931157] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.931491] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.931677] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f7eb624-e1b1-4e17-8b59-c6486cc3817f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.987408] env[61991]: DEBUG oslo_concurrency.lockutils [req-fccec285-1de4-4f30-9b6f-4aef22a1903c req-2a13f0d4-81a4-4dcb-bfdd-5aef04f9e9a7 service nova] Releasing lock "refresh_cache-7ad63ade-cff7-4a32-b567-783db726de55" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.057133] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 947.057400] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 947.057538] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleting the datastore file [datastore1] a21cc5c6-abd8-45c6-b24b-33d3374df65c {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.057815] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5269d8e-9c8e-42cd-90c2-debe6831a383 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.067883] env[61991]: DEBUG oslo_vmware.api [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 947.067883] env[61991]: value = "task-1130140" [ 947.067883] env[61991]: _type = "Task" [ 947.067883] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.078548] env[61991]: DEBUG oslo_vmware.api [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.184920] env[61991]: INFO nova.network.neutron [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Port b603ad92-a23f-4da6-b5dc-60d547ba32ae from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 947.185420] env[61991]: DEBUG nova.network.neutron [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [{"id": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "address": "fa:16:3e:8e:10:f9", "network": {"id": "e6f9d471-4f42-4781-93c3-f4a9ff84317b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-2036205362-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d58eaa6ccc4f46e3a7606f55605d8e15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3f695b6-65bc-45cc-a61d-3c38a14e5c0c", "external-id": "nsx-vlan-transportzone-559", "segmentation_id": 559, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0abb70f-8b", "ovs_interfaceid": "e0abb70f-8b4c-4450-a784-ec0dd2381264", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.204501] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130134, 'name': CreateVM_Task, 'duration_secs': 1.489669} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.205087] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.205827] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.206083] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.206469] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.207054] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f650e3-b113-40d8-b19b-18a1f0607688 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.216294] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130137, 'name': ReconfigVM_Task, 'duration_secs': 0.355707} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.220222] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb10cfe3-d0d8-4b8a-a519-96320d6abeaf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.223135] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfigured VM instance instance-00000054 to attach disk [datastore1] volume-e1d09948-3670-4f76-ad45-f10a94df1f32/volume-e1d09948-3670-4f76-ad45-f10a94df1f32.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.227994] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f11d6d6f-3a00-4a2d-ba54-ae85dd091bbb tempest-ServerTagsTestJSON-1316270002 tempest-ServerTagsTestJSON-1316270002-project-member] Lock "586bf51a-d1c4-44e8-863b-b229a3ca4f2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.281s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.233041] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77ce203d-7cf2-49ee-87c2-a0f54079db77 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.240864] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 947.240864] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df3746-313d-22f2-bcfe-f29fcc2c0033" [ 947.240864] env[61991]: _type = "Task" [ 947.240864] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.267057] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.271753] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f29da06-18df-4baf-a5fe-030972436626 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.274167] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2208c96-e229-447f-9a28-0e7f81bcf854 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.277626] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 947.277626] env[61991]: value = "task-1130141" [ 947.277626] env[61991]: _type = "Task" [ 947.277626] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.282516] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df3746-313d-22f2-bcfe-f29fcc2c0033, 'name': SearchDatastore_Task, 'duration_secs': 0.011042} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.286715] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.287059] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.287355] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.287547] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.287766] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.304808] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2606ab45-3c43-48eb-984c-a94ff7763686 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.307502] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 947.307502] env[61991]: value = "task-1130142" [ 947.307502] env[61991]: _type = "Task" [ 947.307502] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.308246] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.322269] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130141, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.324049] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.324314] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.325493] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3811ee12-6273-4bdb-ae9c-5f8b8070fd6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.331768] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.339251] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 947.339251] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c5fafe-b157-067e-6a22-ba21464ea48b" [ 947.339251] env[61991]: _type = "Task" [ 947.339251] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.348613] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c5fafe-b157-067e-6a22-ba21464ea48b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.419071] env[61991]: DEBUG oslo_vmware.api [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130138, 'name': PowerOffVM_Task, 'duration_secs': 0.218432} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.419400] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 947.419608] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 947.419914] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-193f0b55-ea6a-43a0-adb2-5a68a36a446d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.517890] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 947.518180] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 947.518376] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleting the datastore file [datastore2] 902f158c-fcc4-40a2-abbd-491bdfd4b77a {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.518651] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e32a23c-e869-4aad-ad91-bef124e6fb69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.529169] env[61991]: DEBUG oslo_vmware.api [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 947.529169] env[61991]: value = "task-1130144" [ 947.529169] env[61991]: _type = "Task" [ 947.529169] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.539734] env[61991]: DEBUG oslo_vmware.api [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.550819] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4138fdcb-7a09-494a-9e6a-23ace557dbe2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.559161] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18864725-ef9a-43ed-aab3-665cd710583b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.594495] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3233b68-7620-4f1a-ab40-203ddcd2816f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.603172] env[61991]: DEBUG oslo_vmware.api [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157282} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.605281] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.605543] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 947.605660] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 947.605839] env[61991]: INFO nova.compute.manager [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Took 1.21 seconds to destroy the instance on the hypervisor. [ 947.606135] env[61991]: DEBUG oslo.service.loopingcall [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.606393] env[61991]: DEBUG nova.compute.manager [-] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 947.606493] env[61991]: DEBUG nova.network.neutron [-] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.609117] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7db9bc2-6468-4e2b-81f2-839e9a87654f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.623952] env[61991]: DEBUG nova.compute.provider_tree [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.695017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Releasing lock "refresh_cache-902f158c-fcc4-40a2-abbd-491bdfd4b77a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.792552] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130141, 'name': ReconfigVM_Task, 'duration_secs': 0.228907} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.792864] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 947.794054] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-325b5108-92fc-4440-9f28-4f17737cb643 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.803720] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 947.803720] env[61991]: value = "task-1130145" [ 947.803720] env[61991]: _type = "Task" [ 947.803720] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.812701] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130145, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.827112] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130142, 'name': PowerOffVM_Task, 'duration_secs': 0.219776} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.827385] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 947.827890] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.850813] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c5fafe-b157-067e-6a22-ba21464ea48b, 'name': SearchDatastore_Task, 'duration_secs': 0.010426} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.851687] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d3081e1-c92f-4c8e-a826-e1a94d7ab1ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.858037] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 947.858037] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5221cce4-a5e6-fc65-3314-2b7b376da71c" [ 947.858037] env[61991]: _type = "Task" [ 947.858037] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.866896] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5221cce4-a5e6-fc65-3314-2b7b376da71c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.867710] env[61991]: DEBUG nova.network.neutron [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Port 2dc3450f-efd8-4200-a182-988386f68d00 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 948.042626] env[61991]: DEBUG oslo_vmware.api [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.379215} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.042892] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.043402] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.043637] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.043852] env[61991]: INFO nova.compute.manager [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Took 1.16 seconds to destroy the instance on the hypervisor. [ 948.044149] env[61991]: DEBUG oslo.service.loopingcall [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.044869] env[61991]: DEBUG nova.compute.manager [-] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.044964] env[61991]: DEBUG nova.network.neutron [-] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.056998] env[61991]: DEBUG nova.compute.manager [req-47545f81-4dab-474c-bcbd-714ed120ba21 req-bfb578d3-3d1d-4be6-874f-8facd0f39be1 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Received event network-vif-deleted-0743e42e-3fb0-4809-9305-87b229303bcf {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.057277] env[61991]: INFO nova.compute.manager [req-47545f81-4dab-474c-bcbd-714ed120ba21 req-bfb578d3-3d1d-4be6-874f-8facd0f39be1 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Neutron deleted interface 0743e42e-3fb0-4809-9305-87b229303bcf; detaching it from the instance and deleting it from the info cache [ 948.057545] env[61991]: DEBUG nova.network.neutron [req-47545f81-4dab-474c-bcbd-714ed120ba21 req-bfb578d3-3d1d-4be6-874f-8facd0f39be1 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.127199] env[61991]: DEBUG nova.scheduler.client.report [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.196096] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2be08e32-dc65-472a-b802-6ffdaf7d24c1 tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "interface-902f158c-fcc4-40a2-abbd-491bdfd4b77a-b603ad92-a23f-4da6-b5dc-60d547ba32ae" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.905s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.314366] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130145, 'name': Rename_Task, 'duration_secs': 0.174142} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.314658] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.314902] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cafd7816-b85e-4d86-ab50-02b8b2f14b91 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.321826] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 948.321826] env[61991]: value = "task-1130146" [ 948.321826] env[61991]: _type = "Task" [ 948.321826] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.331644] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130146, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.333423] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 948.333811] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 948.333964] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.334053] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 948.334197] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.334348] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 948.334552] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 948.334714] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 948.334883] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 948.335065] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 948.335250] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 948.340894] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d3df126-3ffd-4c48-9999-22e114d5fd25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.365266] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 948.365266] env[61991]: value = "task-1130147" [ 948.365266] env[61991]: _type = "Task" [ 948.365266] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.378264] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5221cce4-a5e6-fc65-3314-2b7b376da71c, 'name': SearchDatastore_Task, 'duration_secs': 0.011178} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.378264] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.378264] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 948.378557] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6dcddd0a-429d-4af0-98b8-0e92809c1725 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.386549] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130147, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.388057] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 948.388057] env[61991]: value = "task-1130148" [ 948.388057] env[61991]: _type = "Task" [ 948.388057] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.396472] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130148, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.538370] env[61991]: DEBUG nova.network.neutron [-] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.560899] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e73228b-2939-459e-b70e-8521b11eef57 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.571698] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798cd544-fdae-49dd-8e3b-f84fc4444ccc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.613094] env[61991]: DEBUG nova.compute.manager [req-47545f81-4dab-474c-bcbd-714ed120ba21 req-bfb578d3-3d1d-4be6-874f-8facd0f39be1 service nova] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Detach interface failed, port_id=0743e42e-3fb0-4809-9305-87b229303bcf, reason: Instance a21cc5c6-abd8-45c6-b24b-33d3374df65c could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 948.631916] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.949s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.635080] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.622s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.635487] env[61991]: DEBUG nova.objects.instance [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'resources' on Instance uuid 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.658692] env[61991]: INFO nova.scheduler.client.report [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Deleted allocations for instance d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3 [ 948.835989] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130146, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.887482] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130147, 'name': ReconfigVM_Task, 'duration_secs': 0.204563} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.890729] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 948.899566] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.899800] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.899989] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.912860] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130148, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487789} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.912860] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.912860] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.912860] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e8ebb2a-0e11-433c-b41c-08c5f3a095ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.921108] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 948.921108] env[61991]: value = "task-1130149" [ 948.921108] env[61991]: _type = "Task" [ 948.921108] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.936065] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130149, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.044162] env[61991]: INFO nova.compute.manager [-] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Took 1.44 seconds to deallocate network for instance. [ 949.139992] env[61991]: DEBUG nova.objects.instance [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'numa_topology' on Instance uuid 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.168505] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f574b61-3fe3-4714-846c-dddb9c76c65e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.583s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.176533] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 949.177073] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 949.336087] env[61991]: DEBUG oslo_vmware.api [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130146, 'name': PowerOnVM_Task, 'duration_secs': 0.628337} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.336366] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.336588] env[61991]: DEBUG nova.compute.manager [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.337429] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3685752c-b134-434e-ac2b-f58cb6cb6fe5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.401536] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.401809] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.401939] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.402149] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.402359] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.404352] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.404352] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.404352] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.404352] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.404352] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.404352] env[61991]: DEBUG nova.virt.hardware [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.408877] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 949.411055] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fdb1521-ac34-4d98-87d6-409d9e9abb18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.449950] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130149, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080378} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.451488] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.451927] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 949.451927] env[61991]: value = "task-1130150" [ 949.451927] env[61991]: _type = "Task" [ 949.451927] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.452546] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9303f7-c4e1-4f7a-a114-0374522f939e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.464289] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130150, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.482814] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.487951] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46e1e1f4-5ff1-46a1-a4a9-ed6cbed1817d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.508663] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 949.508663] env[61991]: value = "task-1130151" [ 949.508663] env[61991]: _type = "Task" [ 949.508663] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.515990] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.516157] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.516758] env[61991]: DEBUG nova.network.neutron [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 949.523127] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130151, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.549874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.606206] env[61991]: DEBUG nova.network.neutron [-] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.644509] env[61991]: DEBUG nova.objects.base [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Object Instance<56188232-bed0-4a4b-a4bc-01edbb85cbe4> lazy-loaded attributes: resources,numa_topology {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 949.691515] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 949.691761] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 949.859285] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.887749] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.888054] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.888465] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.889061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.889061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.894430] env[61991]: INFO nova.compute.manager [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Terminating instance [ 949.896684] env[61991]: DEBUG nova.compute.manager [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.896793] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.897686] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a6e599-65a9-428e-ba0c-c68beabbf585 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.908761] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.909058] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4292458-5d53-4e86-a021-1f0e6d413237 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.916765] env[61991]: DEBUG oslo_vmware.api [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 949.916765] env[61991]: value = "task-1130152" [ 949.916765] env[61991]: _type = "Task" [ 949.916765] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.930585] env[61991]: DEBUG oslo_vmware.api [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130152, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.967362] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130150, 'name': ReconfigVM_Task, 'duration_secs': 0.407692} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.967598] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 949.969252] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51abdaf-2f56-4377-9e0e-283d76150570 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.972478] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d1fcab-5339-4a95-8392-2b534a549695 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.996404] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf/68268135-4cf4-4732-9104-1720bbb21acf.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.998323] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4be284dc-c494-4c1b-be65-d39f1bedddc7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.012647] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b1d5e9-db24-4782-8743-1d02e63f899e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.029040] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130151, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.054810] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 950.054810] env[61991]: value = "task-1130153" [ 950.054810] env[61991]: _type = "Task" [ 950.054810] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.055983] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d1d22e-b372-479a-af90-7a5a1c0d5fae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.069769] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130153, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.073037] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e125bef8-1fe2-425e-a1fd-1fa7ca964b24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.088087] env[61991]: DEBUG nova.compute.provider_tree [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.092121] env[61991]: DEBUG nova.compute.manager [req-831fba3f-304d-47dc-8c4d-81735b864a66 req-0c8a7a1c-6bb3-4ca6-b37d-5c10d3dbcb6c service nova] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Received event network-vif-deleted-e0abb70f-8b4c-4450-a784-ec0dd2381264 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.113428] env[61991]: INFO nova.compute.manager [-] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Took 2.07 seconds to deallocate network for instance. [ 950.258383] env[61991]: DEBUG nova.network.neutron [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance_info_cache with network_info: [{"id": "2dc3450f-efd8-4200-a182-988386f68d00", "address": "fa:16:3e:de:38:21", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dc3450f-ef", "ovs_interfaceid": "2dc3450f-efd8-4200-a182-988386f68d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.429695] env[61991]: DEBUG oslo_vmware.api [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130152, 'name': PowerOffVM_Task, 'duration_secs': 0.257318} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.429986] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.430180] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.430436] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e50c4aad-ca11-4ec5-ac67-e6a2744e8b09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.502815] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.503093] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.503294] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Deleting the datastore file [datastore2] 7634e00f-48ee-40a0-8625-2f8dc83cd58d {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.503570] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f57dc40-9795-47d6-87cd-1ab2c14d6337 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.511020] env[61991]: DEBUG oslo_vmware.api [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for the task: (returnval){ [ 950.511020] env[61991]: value = "task-1130155" [ 950.511020] env[61991]: _type = "Task" [ 950.511020] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.524172] env[61991]: DEBUG oslo_vmware.api [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130155, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.527555] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130151, 'name': ReconfigVM_Task, 'duration_secs': 0.739822} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.527823] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.528473] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64b21d58-c954-4ad2-bff0-093fee2479eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.535280] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 950.535280] env[61991]: value = "task-1130156" [ 950.535280] env[61991]: _type = "Task" [ 950.535280] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.546655] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130156, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.568921] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130153, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.593508] env[61991]: DEBUG nova.scheduler.client.report [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.619725] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.697166] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Didn't find any instances for network info cache update. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 950.697420] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.697582] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.697731] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.697879] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.698032] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.698239] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.698322] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 950.698475] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.761307] env[61991]: DEBUG oslo_concurrency.lockutils [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.021420] env[61991]: DEBUG oslo_vmware.api [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Task: {'id': task-1130155, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181181} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.021715] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.021881] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.022077] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.022264] env[61991]: INFO nova.compute.manager [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 951.022512] env[61991]: DEBUG oslo.service.loopingcall [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.022710] env[61991]: DEBUG nova.compute.manager [-] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.022807] env[61991]: DEBUG nova.network.neutron [-] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.046248] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130156, 'name': Rename_Task, 'duration_secs': 0.17143} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.046539] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.046783] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c49d842e-3e74-4d79-aa1f-c9b6d756c289 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.054982] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 951.054982] env[61991]: value = "task-1130157" [ 951.054982] env[61991]: _type = "Task" [ 951.054982] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.067680] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130153, 'name': ReconfigVM_Task, 'duration_secs': 0.557787} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.070711] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf/68268135-4cf4-4732-9104-1720bbb21acf.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.071013] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 951.074307] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.098826] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.464s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.102611] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.553s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.102785] env[61991]: DEBUG nova.objects.instance [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lazy-loading 'resources' on Instance uuid a21cc5c6-abd8-45c6-b24b-33d3374df65c {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.203484] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.287420] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2294ae2a-fd7c-4b54-b475-c071958987c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.308155] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99d1d42-0175-4bb2-bd9b-867f45bfb628 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.316909] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 951.479910] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.480224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.566648] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130157, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.577088] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd36700e-f128-424d-bf70-bd22493493be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.597078] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea1523a-45e2-49bc-a80e-31cbb02d88cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.619813] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 951.625461] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b175c051-8eaf-415e-89b2-881fa21eceea tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 30.586s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.626899] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.426s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.627060] env[61991]: INFO nova.compute.manager [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Unshelving [ 951.823283] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.823585] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1302b96e-3093-4409-996b-f32e5f0532aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.832744] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 951.832744] env[61991]: value = "task-1130158" [ 951.832744] env[61991]: _type = "Task" [ 951.832744] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.840827] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.878837] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e330f1-bc5c-44b5-8408-182bab198aa8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.887982] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2105ede3-9364-4715-ab48-945433751712 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.921426] env[61991]: DEBUG nova.network.neutron [-] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.923776] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3256ce-0285-46e3-8222-dda208e8972d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.936127] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e8bce3-6a1d-4f0e-8f17-efec64d7ea4e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.959239] env[61991]: DEBUG nova.compute.provider_tree [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.984277] env[61991]: INFO nova.compute.manager [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Detaching volume e1d09948-3670-4f76-ad45-f10a94df1f32 [ 952.023675] env[61991]: INFO nova.virt.block_device [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Attempting to driver detach volume e1d09948-3670-4f76-ad45-f10a94df1f32 from mountpoint /dev/sdb [ 952.024082] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 952.024162] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 952.025053] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb450169-fe30-4fc4-99af-4294423671af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.047018] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec25d54-8b1a-4b9f-992a-9812dea7c9af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.055317] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e19e20-794d-4c5f-8eac-74a16c810211 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.066092] env[61991]: DEBUG oslo_vmware.api [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130157, 'name': PowerOnVM_Task, 'duration_secs': 0.717957} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.082235] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.082495] env[61991]: INFO nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Took 9.07 seconds to spawn the instance on the hypervisor. [ 952.082686] env[61991]: DEBUG nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.083587] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d27efcb-d32f-48f0-80ce-b9ca68425e83 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.086632] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9d9fb7-ba3e-48ec-9e7a-c3f41ee7e233 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.108305] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] The volume has not been displaced from its original location: [datastore1] volume-e1d09948-3670-4f76-ad45-f10a94df1f32/volume-e1d09948-3670-4f76-ad45-f10a94df1f32.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 952.114035] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfiguring VM instance instance-00000054 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.114300] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c7af0a6-4795-4488-9043-4aae645608a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.131956] env[61991]: DEBUG nova.compute.manager [req-e8afd5f7-2ce9-4ff2-8d8d-91492484b8bf req-0ded6838-3daf-485e-8b68-305de624433e service nova] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Received event network-vif-deleted-490686f3-4e55-4f9f-afe0-6dbb260e6bea {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.142536] env[61991]: DEBUG oslo_vmware.api [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 952.142536] env[61991]: value = "task-1130159" [ 952.142536] env[61991]: _type = "Task" [ 952.142536] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.152097] env[61991]: DEBUG oslo_vmware.api [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130159, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.179799] env[61991]: DEBUG nova.network.neutron [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Port a511534a-b3ea-4087-81d0-c1b3277ecb0f binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 952.344269] env[61991]: DEBUG oslo_vmware.api [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130158, 'name': PowerOnVM_Task, 'duration_secs': 0.473426} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.344477] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.344765] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-34a13677-8b5a-47e0-a2b0-20cd60603999 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance '3b647920-4a69-4d1d-adb1-6fbf716e5514' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 952.424380] env[61991]: INFO nova.compute.manager [-] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Took 1.40 seconds to deallocate network for instance. [ 952.462924] env[61991]: DEBUG nova.scheduler.client.report [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.506192] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.506334] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.621537] env[61991]: INFO nova.compute.manager [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Took 16.22 seconds to build instance. [ 952.655014] env[61991]: DEBUG oslo_vmware.api [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130159, 'name': ReconfigVM_Task, 'duration_secs': 0.475967} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.655341] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Reconfigured VM instance instance-00000054 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 952.660791] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.661039] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7c2d713-834f-4847-8cb2-a4cc39e8d441 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.677588] env[61991]: DEBUG oslo_vmware.api [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 952.677588] env[61991]: value = "task-1130160" [ 952.677588] env[61991]: _type = "Task" [ 952.677588] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.688602] env[61991]: DEBUG oslo_vmware.api [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130160, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.931214] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.967301] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.969682] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.113s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.969877] env[61991]: DEBUG nova.objects.instance [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 952.989906] env[61991]: INFO nova.scheduler.client.report [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleted allocations for instance a21cc5c6-abd8-45c6-b24b-33d3374df65c [ 953.009178] env[61991]: DEBUG nova.compute.utils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.123463] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b988da49-c6bb-4508-9e8c-90caa0f30017 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "7ad63ade-cff7-4a32-b567-783db726de55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.732s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.202283] env[61991]: DEBUG oslo_vmware.api [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130160, 'name': ReconfigVM_Task, 'duration_secs': 0.220901} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.205967] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246936', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'name': 'volume-e1d09948-3670-4f76-ad45-f10a94df1f32', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b3a2ea40-5c4c-4e7d-95b1-38a18c429cec', 'attached_at': '', 'detached_at': '', 'volume_id': 'e1d09948-3670-4f76-ad45-f10a94df1f32', 'serial': 'e1d09948-3670-4f76-ad45-f10a94df1f32'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 953.213806] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "68268135-4cf4-4732-9104-1720bbb21acf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.214100] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.214364] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.498628] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b32d48f0-6f43-4b4d-aa36-c8e5dddb329f tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "a21cc5c6-abd8-45c6-b24b-33d3374df65c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.114s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.511431] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.766477] env[61991]: DEBUG nova.objects.instance [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'flavor' on Instance uuid b3a2ea40-5c4c-4e7d-95b1-38a18c429cec {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.981459] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3c17700b-3432-4bf9-ba04-1c5bb06a0785 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.983118] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.363s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.983118] env[61991]: DEBUG nova.objects.instance [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'resources' on Instance uuid 902f158c-fcc4-40a2-abbd-491bdfd4b77a {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.005248] env[61991]: INFO nova.compute.manager [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Rebuilding instance [ 954.051426] env[61991]: DEBUG nova.compute.manager [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.052405] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0e6001-bd2e-4d45-9c94-900e8634a86f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.269458] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.269727] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.269866] env[61991]: DEBUG nova.network.neutron [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.315933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.316295] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.316527] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.316765] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.316961] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.319568] env[61991]: INFO nova.compute.manager [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Terminating instance [ 954.321378] env[61991]: DEBUG nova.compute.manager [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.321578] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.322420] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a83382-ff07-4f8c-a6c8-3b70536c62f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.331083] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.331383] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc35ce9f-896f-49f3-81c9-a1a7272a53df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.338526] env[61991]: DEBUG oslo_vmware.api [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 954.338526] env[61991]: value = "task-1130161" [ 954.338526] env[61991]: _type = "Task" [ 954.338526] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.347667] env[61991]: DEBUG oslo_vmware.api [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.567385] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.568017] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88f07a91-370c-45b5-a9ea-08fff519461f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.575081] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.575371] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.575621] env[61991]: INFO nova.compute.manager [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Attaching volume 87ec6079-a5ef-48d8-bfef-9127fcf399c1 to /dev/sdb [ 954.582726] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 954.582726] env[61991]: value = "task-1130162" [ 954.582726] env[61991]: _type = "Task" [ 954.582726] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.601750] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.620784] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3946fa30-6f0b-48ef-ad90-0f778f0d465f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.635427] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12002c0d-745e-4a99-b48e-9f0864f37a27 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.655889] env[61991]: DEBUG nova.virt.block_device [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updating existing volume attachment record: a49db7d0-14b8-47a3-ad70-1febd452f0e1 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 954.778135] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca5da37-da8e-4fe2-971b-a88fb18c1a07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.782973] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ecfd0941-85d1-4ff2-9205-a59b1c7e7506 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.303s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.790626] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9036f99f-bb6d-4256-acac-e60c9dece8b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.834364] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d6fc13-b966-482d-af69-d7fe4844bfdd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.850432] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66869840-2dce-4f6c-a47c-f0dc3a3fa90e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.858292] env[61991]: DEBUG oslo_vmware.api [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130161, 'name': PowerOffVM_Task, 'duration_secs': 0.242145} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.859029] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.859251] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.859509] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2091615a-0958-4835-8e0d-d7dce2fc14c3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.869293] env[61991]: DEBUG nova.compute.provider_tree [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.932374] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.932661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.933875] env[61991]: DEBUG nova.compute.manager [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Going to confirm migration 3 {{(pid=61991) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 955.095326] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130162, 'name': PowerOffVM_Task, 'duration_secs': 0.249808} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.095747] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.096112] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.097293] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350d0623-059c-424f-b5f2-2f8c095b2d2f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.106421] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.106710] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-333fdaf7-0842-4ead-a653-fd8c6253f167 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.109473] env[61991]: DEBUG nova.network.neutron [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance_info_cache with network_info: [{"id": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "address": "fa:16:3e:d7:45:df", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa511534a-b3", "ovs_interfaceid": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.372488] env[61991]: DEBUG nova.scheduler.client.report [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.493199] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.493410] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquired lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.493600] env[61991]: DEBUG nova.network.neutron [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.493799] env[61991]: DEBUG nova.objects.instance [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lazy-loading 'info_cache' on Instance uuid 3b647920-4a69-4d1d-adb1-6fbf716e5514 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.612503] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.877586] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.895s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.880138] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.677s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.880334] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.880494] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 955.880860] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.220s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.881013] env[61991]: DEBUG nova.objects.instance [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'pci_requests' on Instance uuid 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.882639] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ab577c-66b6-4573-b4db-87390127ad7f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.889964] env[61991]: DEBUG nova.objects.instance [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'numa_topology' on Instance uuid 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.898143] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ff3c8a-2b6d-4123-a1bd-21900af4ea0f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.904126] env[61991]: INFO nova.scheduler.client.report [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted allocations for instance 902f158c-fcc4-40a2-abbd-491bdfd4b77a [ 955.916483] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccd712d-7286-485d-a9ef-bd73255d9cf7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.923974] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acad341-7b83-4f43-a0d5-931b132a46ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.954370] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179542MB free_disk=167GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 955.954534] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.005112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.005384] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.005593] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.005780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.006026] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.008193] env[61991]: INFO nova.compute.manager [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Terminating instance [ 956.009936] env[61991]: DEBUG nova.compute.manager [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.010233] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.011048] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf27928c-7f4a-4c3a-bcfc-cc6de05595b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.019894] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.020143] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b131c1fa-61eb-4e9e-8e7b-3422aa0517b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.027645] env[61991]: DEBUG oslo_vmware.api [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 956.027645] env[61991]: value = "task-1130166" [ 956.027645] env[61991]: _type = "Task" [ 956.027645] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.035641] env[61991]: DEBUG oslo_vmware.api [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130166, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.135209] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5ac7a8-1e6c-4363-b3e7-0836263ae1a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.157832] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068fc869-8a6e-4610-9b69-cc32d5987f5f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.166227] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 956.392564] env[61991]: INFO nova.compute.claims [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.419552] env[61991]: DEBUG oslo_concurrency.lockutils [None req-daa13a4f-fd46-4b64-94b7-2bf2c034760b tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "902f158c-fcc4-40a2-abbd-491bdfd4b77a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.541s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.536250] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.536500] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.541062] env[61991]: DEBUG oslo_vmware.api [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130166, 'name': PowerOffVM_Task, 'duration_secs': 0.191466} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.541440] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.541566] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.541804] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57d78165-e142-4098-bfe4-2a7679ecbff1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.671941] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.672250] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6a55438-6737-4376-b13f-dea7d5aabf1f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.681678] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 956.681678] env[61991]: value = "task-1130168" [ 956.681678] env[61991]: _type = "Task" [ 956.681678] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.689067] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.702314] env[61991]: DEBUG nova.network.neutron [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance_info_cache with network_info: [{"id": "2dc3450f-efd8-4200-a182-988386f68d00", "address": "fa:16:3e:de:38:21", "network": {"id": "f6e53dd7-b54c-4eda-81e5-0ecbf369c9ee", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1575599144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4a18b2a73f7643e1b89c9660a770e9da", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e5d88cd9-35a3-4ac3-9d6d-756464cd6cc5", "external-id": "nsx-vlan-transportzone-685", "segmentation_id": 685, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2dc3450f-ef", "ovs_interfaceid": "2dc3450f-efd8-4200-a182-988386f68d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.720487] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.720739] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.721068] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleting the datastore file [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.721263] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.721446] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.721612] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleting the datastore file [datastore1] 69a272a7-8da6-4fb2-95be-85cfb7d70d44 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.721851] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1b3ffa9-9cb7-4655-8089-b3b3e31993e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.724120] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ab52f96-83bd-4f66-8566-d36b90e6e37e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.732104] env[61991]: DEBUG oslo_vmware.api [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for the task: (returnval){ [ 956.732104] env[61991]: value = "task-1130170" [ 956.732104] env[61991]: _type = "Task" [ 956.732104] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.733806] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 956.733806] env[61991]: value = "task-1130169" [ 956.733806] env[61991]: _type = "Task" [ 956.733806] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.744758] env[61991]: DEBUG oslo_vmware.api [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.747709] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.042461] env[61991]: DEBUG nova.compute.utils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.192763] env[61991]: DEBUG oslo_vmware.api [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130168, 'name': PowerOnVM_Task, 'duration_secs': 0.410782} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.193127] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.193265] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8640faad-f7e4-4b07-9d27-0dbb6d643ab8 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance '68268135-4cf4-4732-9104-1720bbb21acf' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.204781] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Releasing lock "refresh_cache-3b647920-4a69-4d1d-adb1-6fbf716e5514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.205027] env[61991]: DEBUG nova.objects.instance [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lazy-loading 'migration_context' on Instance uuid 3b647920-4a69-4d1d-adb1-6fbf716e5514 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.246436] env[61991]: DEBUG oslo_vmware.api [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Task: {'id': task-1130170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168313} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.249344] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.249538] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.249724] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.249899] env[61991]: INFO nova.compute.manager [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Took 2.93 seconds to destroy the instance on the hypervisor. [ 957.250152] env[61991]: DEBUG oslo.service.loopingcall [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.250358] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165779} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.250554] env[61991]: DEBUG nova.compute.manager [-] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.250649] env[61991]: DEBUG nova.network.neutron [-] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 957.252463] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.252647] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.252825] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.545616] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.591291] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e07217-cc73-4a13-9751-fb5b04663eee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.599148] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a833a78c-9471-4423-aa0c-b84b22987439 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.628432] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8f9616-7e35-4258-bd30-b8369e162c09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.635418] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ca780c-19ff-4c20-9c9a-9239f178431c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.648875] env[61991]: DEBUG nova.compute.provider_tree [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.708071] env[61991]: DEBUG nova.objects.base [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Object Instance<3b647920-4a69-4d1d-adb1-6fbf716e5514> lazy-loaded attributes: info_cache,migration_context {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 957.708917] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae335845-86a7-4eb0-a123-599914ed825a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.729079] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23852ae8-ece9-467c-adde-79d4c9f2f687 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.734778] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 957.734778] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284b898-1eb3-beff-9254-d45bbf79791a" [ 957.734778] env[61991]: _type = "Task" [ 957.734778] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.744686] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284b898-1eb3-beff-9254-d45bbf79791a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.963511] env[61991]: DEBUG nova.network.neutron [-] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.151457] env[61991]: DEBUG nova.scheduler.client.report [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.247958] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284b898-1eb3-beff-9254-d45bbf79791a, 'name': SearchDatastore_Task, 'duration_secs': 0.007615} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.249011] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.285690] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.285952] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.286160] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.286363] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.286514] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.286665] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.286872] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.287060] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.287259] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.287431] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.287602] env[61991]: DEBUG nova.virt.hardware [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.288524] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b82a5c-cdd5-4d93-8089-f92520bf6988 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.296671] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ceae34-1089-4d24-8490-f84e1c0c7dc6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.311364] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:f3:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f039d7a6-60dd-49c4-95bc-ba71796b1aba', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.323454] env[61991]: DEBUG oslo.service.loopingcall [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.324520] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.324625] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6922d244-8162-4aaf-ae6f-fcf3bd8e4965 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.344523] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.344523] env[61991]: value = "task-1130172" [ 958.344523] env[61991]: _type = "Task" [ 958.344523] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.352634] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130172, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.410010] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.410208] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.410587] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore1] b3a2ea40-5c4c-4e7d-95b1-38a18c429cec {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.411661] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-034c7234-b308-4f46-b549-e1386f0e7130 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.420069] env[61991]: DEBUG oslo_vmware.api [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 958.420069] env[61991]: value = "task-1130173" [ 958.420069] env[61991]: _type = "Task" [ 958.420069] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.432029] env[61991]: DEBUG oslo_vmware.api [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.467115] env[61991]: INFO nova.compute.manager [-] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Took 1.22 seconds to deallocate network for instance. [ 958.544527] env[61991]: DEBUG nova.compute.manager [req-05921ec3-5771-4e2f-9aaa-e4d8475dcd66 req-1c639db7-a80e-4033-9b84-840fe37e9d5a service nova] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Received event network-vif-deleted-9b9cd766-2c3d-4f0a-beb3-30531d95c2ff {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.610600] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.610985] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.611142] env[61991]: INFO nova.compute.manager [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Attaching volume 3d090d84-cb6e-4050-b26a-89d77c479ef4 to /dev/sdb [ 958.644339] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0393408-b592-486b-b07c-2d0616892280 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.652721] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd19475-f221-4648-ac58-d6d6b97ff2de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.656244] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.775s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.658630] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.728s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.658877] env[61991]: DEBUG nova.objects.instance [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lazy-loading 'resources' on Instance uuid 7634e00f-48ee-40a0-8625-2f8dc83cd58d {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.670909] env[61991]: DEBUG nova.virt.block_device [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating existing volume attachment record: 94bbafde-0ca7-4373-bcd0-3b988d676f75 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 958.689736] env[61991]: INFO nova.network.neutron [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating port c7263e46-a164-4813-be3e-faae7fc31ff2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 958.854612] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130172, 'name': CreateVM_Task, 'duration_secs': 0.351134} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.854784] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.855494] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.855932] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.856310] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.856576] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-484553f4-82a5-40a8-9273-1e3282a1fa6b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.861902] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 958.861902] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524aefdc-be76-89ff-cacf-12ea7ee95aae" [ 958.861902] env[61991]: _type = "Task" [ 958.861902] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.870084] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524aefdc-be76-89ff-cacf-12ea7ee95aae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.930435] env[61991]: DEBUG oslo_vmware.api [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130173, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175463} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.930688] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.930877] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.931067] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.931256] env[61991]: INFO nova.compute.manager [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Took 2.92 seconds to destroy the instance on the hypervisor. [ 958.931501] env[61991]: DEBUG oslo.service.loopingcall [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.931687] env[61991]: DEBUG nova.compute.manager [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.931785] env[61991]: DEBUG nova.network.neutron [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 958.974568] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.203701] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.204019] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.204275] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.204946] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.204946] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.206896] env[61991]: INFO nova.compute.manager [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Terminating instance [ 959.209264] env[61991]: DEBUG nova.compute.manager [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.209691] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.210636] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c49217-afa5-4bed-9d0c-dfd971a665e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.216586] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 959.216798] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246941', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'name': 'volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0745dd19-6b08-447a-8333-047aeaa1fa83', 'attached_at': '', 'detached_at': '', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'serial': '87ec6079-a5ef-48d8-bfef-9127fcf399c1'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 959.217814] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eede79c2-df8e-4985-8502-00f62615cebc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.238040] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.240906] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6835a5e7-5811-4416-a9eb-e66e8a102a63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.244423] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71985be2-66be-48bc-ae91-d69a0a9e42a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.263541] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "68268135-4cf4-4732-9104-1720bbb21acf" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.264116] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.264387] env[61991]: DEBUG nova.compute.manager [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Going to confirm migration 4 {{(pid=61991) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 959.273090] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1/volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.277614] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab1e2084-c9a4-44dd-99d1-091fa064210d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.291860] env[61991]: DEBUG oslo_vmware.api [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 959.291860] env[61991]: value = "task-1130177" [ 959.291860] env[61991]: _type = "Task" [ 959.291860] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.303025] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 959.303025] env[61991]: value = "task-1130178" [ 959.303025] env[61991]: _type = "Task" [ 959.303025] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.310559] env[61991]: DEBUG oslo_vmware.api [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.320810] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130178, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.374109] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524aefdc-be76-89ff-cacf-12ea7ee95aae, 'name': SearchDatastore_Task, 'duration_secs': 0.009906} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.374438] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.374697] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.374909] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.375078] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.375286] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.375885] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00d95807-4b91-45d3-9cf0-8964a3b6b087 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.386200] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.386423] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.389671] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a91e241-aaef-421e-bf33-233fd9905478 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.396901] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 959.396901] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d8a09f-8ac6-4aad-69f2-ace29dd69888" [ 959.396901] env[61991]: _type = "Task" [ 959.396901] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.408737] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d8a09f-8ac6-4aad-69f2-ace29dd69888, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.488200] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b647f377-45be-46bb-9932-9a316d88c05e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.497529] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebd7e0c-76b6-4d89-9711-77f4b9fad034 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.529295] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85ba828-226a-40c4-8f0a-57ebdc4526e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.538049] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232004de-db42-47a7-af6f-eed6c89519cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.552565] env[61991]: DEBUG nova.compute.provider_tree [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 959.569857] env[61991]: DEBUG nova.compute.manager [req-a952c0b3-2783-4340-a3de-501c30826127 req-c9c8b6aa-26ac-42a9-929d-36a721271b23 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Received event network-vif-deleted-c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.570046] env[61991]: INFO nova.compute.manager [req-a952c0b3-2783-4340-a3de-501c30826127 req-c9c8b6aa-26ac-42a9-929d-36a721271b23 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Neutron deleted interface c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5; detaching it from the instance and deleting it from the info cache [ 959.570236] env[61991]: DEBUG nova.network.neutron [req-a952c0b3-2783-4340-a3de-501c30826127 req-c9c8b6aa-26ac-42a9-929d-36a721271b23 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.803280] env[61991]: DEBUG oslo_vmware.api [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130177, 'name': PowerOffVM_Task, 'duration_secs': 0.237024} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.803590] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.803736] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.804066] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-089755c8-3a6b-4743-a90b-e1857821fbca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.814467] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130178, 'name': ReconfigVM_Task, 'duration_secs': 0.506977} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.814755] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1/volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.819545] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96d8de58-00e4-4d62-9c47-bb61b7c6d943 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.836796] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 959.836796] env[61991]: value = "task-1130180" [ 959.836796] env[61991]: _type = "Task" [ 959.836796] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.845715] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130180, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.855030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.855243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.855427] env[61991]: DEBUG nova.network.neutron [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 959.855618] env[61991]: DEBUG nova.objects.instance [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lazy-loading 'info_cache' on Instance uuid 68268135-4cf4-4732-9104-1720bbb21acf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.868421] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.868421] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.868672] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleting the datastore file [datastore1] 9d1ac179-3d7e-4f0b-a80d-033eba447cab {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.868871] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2107d551-e64c-431a-b511-3c09680aa7b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.876098] env[61991]: DEBUG oslo_vmware.api [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for the task: (returnval){ [ 959.876098] env[61991]: value = "task-1130181" [ 959.876098] env[61991]: _type = "Task" [ 959.876098] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.885543] env[61991]: DEBUG oslo_vmware.api [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.908377] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d8a09f-8ac6-4aad-69f2-ace29dd69888, 'name': SearchDatastore_Task, 'duration_secs': 0.010191} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.909330] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5ffe6ff-7613-4444-b880-fe46a25e263f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.920457] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 959.920457] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521696bf-9c65-8d3f-de13-88dc2608130b" [ 959.920457] env[61991]: _type = "Task" [ 959.920457] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.932719] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521696bf-9c65-8d3f-de13-88dc2608130b, 'name': SearchDatastore_Task, 'duration_secs': 0.013004} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.933336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.933674] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.934038] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-994a3b65-682e-467e-a2b2-bdbd42390573 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.942367] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 959.942367] env[61991]: value = "task-1130182" [ 959.942367] env[61991]: _type = "Task" [ 959.942367] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.952546] env[61991]: DEBUG nova.network.neutron [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.954340] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130182, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.073031] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49c9e968-859b-4ee0-9798-97fc5d7e2f5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.083563] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fcb702-e20a-44ff-b3b6-975b7ce613eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.095606] env[61991]: ERROR nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] [req-76f80b66-971e-43b4-8d76-52ce2a318f84] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-76f80b66-971e-43b4-8d76-52ce2a318f84"}]} [ 960.124512] env[61991]: DEBUG nova.compute.manager [req-a952c0b3-2783-4340-a3de-501c30826127 req-c9c8b6aa-26ac-42a9-929d-36a721271b23 service nova] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Detach interface failed, port_id=c2cb6230-aa1a-4cf2-950e-5133d7a2b1c5, reason: Instance b3a2ea40-5c4c-4e7d-95b1-38a18c429cec could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 960.125718] env[61991]: DEBUG nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 960.143607] env[61991]: DEBUG nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 960.143951] env[61991]: DEBUG nova.compute.provider_tree [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 960.159486] env[61991]: DEBUG nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 960.180979] env[61991]: DEBUG nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 960.350673] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130180, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.387434] env[61991]: DEBUG oslo_vmware.api [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Task: {'id': task-1130181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10866} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.387710] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.387904] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.388102] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.388382] env[61991]: INFO nova.compute.manager [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Took 1.18 seconds to destroy the instance on the hypervisor. [ 960.388665] env[61991]: DEBUG oslo.service.loopingcall [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.389211] env[61991]: DEBUG nova.compute.manager [-] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.389314] env[61991]: DEBUG nova.network.neutron [-] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.417278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.417278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.417366] env[61991]: DEBUG nova.network.neutron [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.455055] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130182, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.455521] env[61991]: INFO nova.compute.manager [-] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Took 1.52 seconds to deallocate network for instance. [ 960.461233] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b876b5-b0a3-46a3-9a7b-414147dbde61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.469520] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8027f5ed-af2f-41d9-a8be-4a0e22f8f9d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.501374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3f4379-9e33-4d40-9cdb-c764a5ac0cca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.509738] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83536808-4d62-4b29-9262-dfd75708d2f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.523688] env[61991]: DEBUG nova.compute.provider_tree [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 960.571212] env[61991]: DEBUG nova.compute.manager [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-vif-plugged-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.571437] env[61991]: DEBUG oslo_concurrency.lockutils [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.571655] env[61991]: DEBUG oslo_concurrency.lockutils [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.571827] env[61991]: DEBUG oslo_concurrency.lockutils [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.572021] env[61991]: DEBUG nova.compute.manager [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] No waiting events found dispatching network-vif-plugged-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.572191] env[61991]: WARNING nova.compute.manager [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received unexpected event network-vif-plugged-c7263e46-a164-4813-be3e-faae7fc31ff2 for instance with vm_state shelved_offloaded and task_state spawning. [ 960.572369] env[61991]: DEBUG nova.compute.manager [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-changed-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.572569] env[61991]: DEBUG nova.compute.manager [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Refreshing instance network info cache due to event network-changed-c7263e46-a164-4813-be3e-faae7fc31ff2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.572693] env[61991]: DEBUG oslo_concurrency.lockutils [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.849970] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130180, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.001696] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130182, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530921} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.001696] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.001696] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.001696] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-abcedace-867d-4fa3-8372-451065397eae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.001696] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.001696] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 961.001696] env[61991]: value = "task-1130183" [ 961.001696] env[61991]: _type = "Task" [ 961.001696] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.001696] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.129813] env[61991]: DEBUG nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 961.130205] env[61991]: DEBUG nova.compute.provider_tree [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 131 to 132 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 961.130318] env[61991]: DEBUG nova.compute.provider_tree [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.140025] env[61991]: DEBUG nova.network.neutron [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance_info_cache with network_info: [{"id": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "address": "fa:16:3e:d7:45:df", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa511534a-b3", "ovs_interfaceid": "a511534a-b3ea-4087-81d0-c1b3277ecb0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.143425] env[61991]: DEBUG nova.network.neutron [-] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.218544] env[61991]: DEBUG nova.network.neutron [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.351687] env[61991]: DEBUG oslo_vmware.api [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130180, 'name': ReconfigVM_Task, 'duration_secs': 1.170248} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.352021] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246941', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'name': 'volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0745dd19-6b08-447a-8333-047aeaa1fa83', 'attached_at': '', 'detached_at': '', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'serial': '87ec6079-a5ef-48d8-bfef-9127fcf399c1'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 961.482599] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075308} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.482884] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.483688] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34990aa-3775-4598-88a2-148fb6a70f90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.505691] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.506034] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8bde60c4-ef04-4c6e-b1c5-f032d86a7537 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.526167] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 961.526167] env[61991]: value = "task-1130185" [ 961.526167] env[61991]: _type = "Task" [ 961.526167] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.534267] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.599089] env[61991]: DEBUG nova.compute.manager [req-891ccb05-d6e0-439d-a18f-d75fd236fdec req-764d298b-e18b-47d2-a3d8-148537c372fd service nova] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Received event network-vif-deleted-f07a2552-10f4-48c4-9a79-b66506f34302 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.637398] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.979s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.640064] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.685s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.643947] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-68268135-4cf4-4732-9104-1720bbb21acf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.644215] env[61991]: DEBUG nova.objects.instance [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lazy-loading 'migration_context' on Instance uuid 68268135-4cf4-4732-9104-1720bbb21acf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.645244] env[61991]: INFO nova.compute.manager [-] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Took 1.26 seconds to deallocate network for instance. [ 961.721098] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.724024] env[61991]: DEBUG oslo_concurrency.lockutils [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.724024] env[61991]: DEBUG nova.network.neutron [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Refreshing network info cache for port c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.752358] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='40d753afa4014a1ae852052e63f12ffe',container_format='bare',created_at=2024-10-04T09:24:10Z,direct_url=,disk_format='vmdk',id=5cf3221b-0c05-4b11-a46c-192b39b794ad,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1043247066-shelved',owner='9a2b37cc1ade437097583e823d457ca6',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-04T09:24:27Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.752633] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.752800] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.752980] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.753149] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.753304] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.753510] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.753673] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.753843] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.754045] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.754260] env[61991]: DEBUG nova.virt.hardware [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.755140] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8aaf48-b4f6-46e9-80f9-1d49a569baee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.765841] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c235215a-d6d4-4a5e-be42-4eb9a0099f53 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.780559] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:d4:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7263e46-a164-4813-be3e-faae7fc31ff2', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.788065] env[61991]: DEBUG oslo.service.loopingcall [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.788065] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.788202] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3743ff5-17b0-41b7-aa50-e6cef8299072 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.808328] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.808328] env[61991]: value = "task-1130186" [ 961.808328] env[61991]: _type = "Task" [ 961.808328] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.817823] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130186, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.037755] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130185, 'name': ReconfigVM_Task, 'duration_secs': 0.307559} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.038142] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55/7ad63ade-cff7-4a32-b567-783db726de55.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.038800] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-453fb7db-bf5f-4460-b793-f0f1a0b765f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.046934] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 962.046934] env[61991]: value = "task-1130187" [ 962.046934] env[61991]: _type = "Task" [ 962.046934] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.056302] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130187, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.150223] env[61991]: DEBUG nova.objects.base [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Object Instance<68268135-4cf4-4732-9104-1720bbb21acf> lazy-loaded attributes: info_cache,migration_context {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 962.151270] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a2bd5a-ed3c-4d96-9796-ed063d91e26f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.170843] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7165860c-bf28-4610-8f1a-cdc79db84abd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.176710] env[61991]: DEBUG oslo_vmware.api [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 962.176710] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52827a9f-faf5-6f12-63c5-8d9d714cfef9" [ 962.176710] env[61991]: _type = "Task" [ 962.176710] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.185453] env[61991]: DEBUG oslo_vmware.api [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52827a9f-faf5-6f12-63c5-8d9d714cfef9, 'name': SearchDatastore_Task, 'duration_secs': 0.007141} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.185707] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.320314] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130186, 'name': CreateVM_Task, 'duration_secs': 0.341475} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.320487] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.321146] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.321319] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.321841] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.322108] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4713769-c6b4-4c4f-974d-2bc5d9efd126 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.327367] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 962.327367] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520a5aea-6c67-0edd-c6a3-0c0d8e364af1" [ 962.327367] env[61991]: _type = "Task" [ 962.327367] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.339392] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520a5aea-6c67-0edd-c6a3-0c0d8e364af1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.421925] env[61991]: DEBUG nova.network.neutron [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updated VIF entry in instance network info cache for port c7263e46-a164-4813-be3e-faae7fc31ff2. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.422396] env[61991]: DEBUG nova.network.neutron [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.557981] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130187, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.655229] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Applying migration context for instance 3b647920-4a69-4d1d-adb1-6fbf716e5514 as it has an incoming, in-progress migration f5a2581b-5e86-49df-9a62-2b362ef342c1. Migration status is finished {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 962.655492] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Applying migration context for instance 68268135-4cf4-4732-9104-1720bbb21acf as it has an incoming, in-progress migration ac7f2f11-e6d1-4f68-81ba-58a9c7b870d1. Migration status is confirming {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 962.657281] env[61991]: INFO nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating resource usage from migration f5a2581b-5e86-49df-9a62-2b362ef342c1 [ 962.657595] env[61991]: INFO nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating resource usage from migration ac7f2f11-e6d1-4f68-81ba-58a9c7b870d1 [ 962.673347] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.673487] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance d36bd981-9ea3-46f6-8376-ac1e0c3bf61e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.673610] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.673727] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f7eab1a9-2a8a-4664-8736-ca599e207c31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.673839] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9d1ac179-3d7e-4f0b-a80d-033eba447cab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.673972] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance b3a2ea40-5c4c-4e7d-95b1-38a18c429cec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 962.674141] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 69a272a7-8da6-4fb2-95be-85cfb7d70d44 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 962.674264] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 0745dd19-6b08-447a-8333-047aeaa1fa83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.674376] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 5364f93c-ca53-480b-8397-c331dfdeed3f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.674486] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 1730f9b3-69ad-476b-b3ab-3e1345f2a115 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.674594] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Migration f5a2581b-5e86-49df-9a62-2b362ef342c1 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 962.674702] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 3b647920-4a69-4d1d-adb1-6fbf716e5514 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.674810] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 7ad63ade-cff7-4a32-b567-783db726de55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.674918] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Migration ac7f2f11-e6d1-4f68-81ba-58a9c7b870d1 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 962.675036] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 68268135-4cf4-4732-9104-1720bbb21acf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.675147] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 56188232-bed0-4a4b-a4bc-01edbb85cbe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 962.675336] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 962.675465] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3328MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 962.829954] env[61991]: INFO nova.scheduler.client.report [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Deleted allocations for instance 7634e00f-48ee-40a0-8625-2f8dc83cd58d [ 962.850257] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.850528] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Processing image 5cf3221b-0c05-4b11-a46c-192b39b794ad {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.850774] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.851456] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.851456] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.853041] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a82b342d-08de-4a14-b265-3cb951bcae4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.856318] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae07599-decb-48fe-a0e0-e4273d3cb9b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.865294] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f091a8-ba03-4192-9eac-b294bfb6ec14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.869548] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.869810] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.870926] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3af7b518-1219-4c38-9fc0-fe817669138b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.915904] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd2df00-3b12-4760-bad5-e422fbcf764f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.919215] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 962.919215] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5265bc50-db14-d10d-a7fb-3da07879cb73" [ 962.919215] env[61991]: _type = "Task" [ 962.919215] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.925041] env[61991]: DEBUG oslo_concurrency.lockutils [req-f3bf8dee-8dfd-466d-8fd7-47651c6d86d5 req-04d245de-b28e-438c-8b1f-b92d748bdfe2 service nova] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.926906] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59fe644-24f8-4178-b189-3e992cecd34b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.934720] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Preparing fetch location {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 962.934972] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Fetch image to [datastore1] OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad/OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad.vmdk {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 962.935173] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Downloading stream optimized image 5cf3221b-0c05-4b11-a46c-192b39b794ad to [datastore1] OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad/OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad.vmdk on the data store datastore1 as vApp {{(pid=61991) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 962.935350] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Downloading image file data 5cf3221b-0c05-4b11-a46c-192b39b794ad to the ESX as VM named 'OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad' {{(pid=61991) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 962.944979] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.005339] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 963.005339] env[61991]: value = "resgroup-9" [ 963.005339] env[61991]: _type = "ResourcePool" [ 963.005339] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 963.005610] env[61991]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-feadd9ce-5305-4ebe-8214-eed7a6b59238 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.028285] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lease: (returnval){ [ 963.028285] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521bdc8f-9e6b-5d91-a5ad-dfaa21d580b6" [ 963.028285] env[61991]: _type = "HttpNfcLease" [ 963.028285] env[61991]: } obtained for vApp import into resource pool (val){ [ 963.028285] env[61991]: value = "resgroup-9" [ 963.028285] env[61991]: _type = "ResourcePool" [ 963.028285] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 963.028623] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the lease: (returnval){ [ 963.028623] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521bdc8f-9e6b-5d91-a5ad-dfaa21d580b6" [ 963.028623] env[61991]: _type = "HttpNfcLease" [ 963.028623] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 963.035345] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.035345] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521bdc8f-9e6b-5d91-a5ad-dfaa21d580b6" [ 963.035345] env[61991]: _type = "HttpNfcLease" [ 963.035345] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 963.061467] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130187, 'name': Rename_Task, 'duration_secs': 0.872674} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.061847] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.062117] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fe4cc2d-1697-4e3f-98ec-0a2ae9e16b2d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.070276] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 963.070276] env[61991]: value = "task-1130189" [ 963.070276] env[61991]: _type = "Task" [ 963.070276] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.078992] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130189, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.153079] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.342101] env[61991]: DEBUG nova.objects.instance [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'flavor' on Instance uuid 0745dd19-6b08-447a-8333-047aeaa1fa83 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.344306] env[61991]: DEBUG oslo_concurrency.lockutils [None req-dc1a29e6-1672-431c-824b-5f3df0ccc44e tempest-ListImageFiltersTestJSON-526147395 tempest-ListImageFiltersTestJSON-526147395-project-member] Lock "7634e00f-48ee-40a0-8625-2f8dc83cd58d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.456s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.448105] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.537447] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.537447] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521bdc8f-9e6b-5d91-a5ad-dfaa21d580b6" [ 963.537447] env[61991]: _type = "HttpNfcLease" [ 963.537447] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 963.537790] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 963.537790] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]521bdc8f-9e6b-5d91-a5ad-dfaa21d580b6" [ 963.537790] env[61991]: _type = "HttpNfcLease" [ 963.537790] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 963.538535] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ea5866-7065-44b6-a7a6-69b1b057e1a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.546734] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c857a8-957c-b681-7bbd-3f8dd47c0aa0/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 963.547433] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c857a8-957c-b681-7bbd-3f8dd47c0aa0/disk-0.vmdk. {{(pid=61991) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 963.616687] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1910e6a9-cf3a-4147-8bb8-270b68e186a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.618654] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130189, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.722485] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 963.722485] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246944', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'name': 'volume-3d090d84-cb6e-4050-b26a-89d77c479ef4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5364f93c-ca53-480b-8397-c331dfdeed3f', 'attached_at': '', 'detached_at': '', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'serial': '3d090d84-cb6e-4050-b26a-89d77c479ef4'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 963.723300] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a1e77b-b7ec-4eed-b8a3-457a53d4e064 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.742816] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59760092-9d44-491e-99a1-e75b5dea44b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.762096] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.770211] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] volume-3d090d84-cb6e-4050-b26a-89d77c479ef4/volume-3d090d84-cb6e-4050-b26a-89d77c479ef4.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.773552] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-243fb688-9628-4d7a-a322-f4141fb994a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.793839] env[61991]: DEBUG oslo_vmware.api [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 963.793839] env[61991]: value = "task-1130190" [ 963.793839] env[61991]: _type = "Task" [ 963.793839] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.803526] env[61991]: DEBUG oslo_vmware.api [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130190, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.848947] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7a5eab22-e14f-4a91-a9bd-5839fd47d9c1 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.273s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.850188] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.088s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.953800] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 963.954044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.314s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.954350] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.705s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.118273] env[61991]: DEBUG oslo_vmware.api [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130189, 'name': PowerOnVM_Task, 'duration_secs': 0.696204} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.118273] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.118273] env[61991]: DEBUG nova.compute.manager [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.118709] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4bc90f-7342-4001-850e-ec05eb2bfdad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.307566] env[61991]: DEBUG oslo_vmware.api [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130190, 'name': ReconfigVM_Task, 'duration_secs': 0.440228} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.309585] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfigured VM instance instance-0000005c to attach disk [datastore1] volume-3d090d84-cb6e-4050-b26a-89d77c479ef4/volume-3d090d84-cb6e-4050-b26a-89d77c479ef4.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.315622] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d250a5ff-6706-4c71-81b3-8a2eb8247a8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.336175] env[61991]: DEBUG oslo_vmware.api [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 964.336175] env[61991]: value = "task-1130191" [ 964.336175] env[61991]: _type = "Task" [ 964.336175] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.347087] env[61991]: DEBUG oslo_vmware.api [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.357020] env[61991]: INFO nova.compute.manager [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Detaching volume 87ec6079-a5ef-48d8-bfef-9127fcf399c1 [ 964.401622] env[61991]: INFO nova.virt.block_device [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Attempting to driver detach volume 87ec6079-a5ef-48d8-bfef-9127fcf399c1 from mountpoint /dev/sdb [ 964.401712] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 964.401871] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246941', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'name': 'volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0745dd19-6b08-447a-8333-047aeaa1fa83', 'attached_at': '', 'detached_at': '', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'serial': '87ec6079-a5ef-48d8-bfef-9127fcf399c1'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 964.402790] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e424d1a4-cd1f-45dd-a9d0-9bb4c622d1c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.426321] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8547b80b-a4d9-479d-a270-730e28823d30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.436057] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01c0195-f6ee-4a5c-a2ad-104efea3806e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.470089] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9944c611-d01b-4687-84fc-3b4823048b44 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.495494] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] The volume has not been displaced from its original location: [datastore2] volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1/volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 964.503327] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 964.516067] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5654a28a-eea9-423c-9958-533173d71185 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.544922] env[61991]: DEBUG oslo_vmware.api [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 964.544922] env[61991]: value = "task-1130192" [ 964.544922] env[61991]: _type = "Task" [ 964.544922] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.562392] env[61991]: DEBUG oslo_vmware.api [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130192, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.643572] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.762899] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a25480-2e86-4185-8ea4-815dfa292e21 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.771732] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfeb30c-7f1a-4a66-9903-c9caf31c6fd7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.808441] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98eb48ce-1b74-45b2-b904-6a0fb30f6183 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.817700] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4892f06-b2ef-4146-81a2-2b65793e9455 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.834265] env[61991]: DEBUG nova.compute.provider_tree [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.849114] env[61991]: DEBUG oslo_vmware.api [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130191, 'name': ReconfigVM_Task, 'duration_secs': 0.166257} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.849537] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246944', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'name': 'volume-3d090d84-cb6e-4050-b26a-89d77c479ef4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5364f93c-ca53-480b-8397-c331dfdeed3f', 'attached_at': '', 'detached_at': '', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'serial': '3d090d84-cb6e-4050-b26a-89d77c479ef4'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 964.916039] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Completed reading data from the image iterator. {{(pid=61991) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 964.916039] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c857a8-957c-b681-7bbd-3f8dd47c0aa0/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 964.916039] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38956675-3fe8-4e4c-b484-d9d98f8cd171 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.923439] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c857a8-957c-b681-7bbd-3f8dd47c0aa0/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 964.923815] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c857a8-957c-b681-7bbd-3f8dd47c0aa0/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 964.924185] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7ba39cda-1bdd-43e5-9f5a-c4dc278a0ba0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.055871] env[61991]: DEBUG oslo_vmware.api [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130192, 'name': ReconfigVM_Task, 'duration_secs': 0.30854} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.056249] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 965.061077] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12c7c43a-00af-420e-a01c-56d8df8d5077 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.078300] env[61991]: DEBUG oslo_vmware.api [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 965.078300] env[61991]: value = "task-1130193" [ 965.078300] env[61991]: _type = "Task" [ 965.078300] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.087291] env[61991]: DEBUG oslo_vmware.api [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130193, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.321402] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c857a8-957c-b681-7bbd-3f8dd47c0aa0/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 965.321688] env[61991]: INFO nova.virt.vmwareapi.images [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Downloaded image file data 5cf3221b-0c05-4b11-a46c-192b39b794ad [ 965.322779] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4436e179-9f9d-477a-8436-0b5215fe17a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.341499] env[61991]: DEBUG nova.scheduler.client.report [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.346406] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16240dcd-1d36-47d0-a801-1978fb7e8111 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.425340] env[61991]: INFO nova.virt.vmwareapi.images [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] The imported VM was unregistered [ 965.428213] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Caching image {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 965.428524] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Creating directory with path [datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.429184] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-051c5584-0ddb-4899-a97c-ada23233f5e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.459983] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Created directory with path [datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.460270] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad/OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad.vmdk to [datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk. {{(pid=61991) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 965.460612] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-58cb23cb-a7c0-456b-a39e-68b1ff6ddb4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.472421] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 965.472421] env[61991]: value = "task-1130195" [ 965.472421] env[61991]: _type = "Task" [ 965.472421] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.483284] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.590274] env[61991]: DEBUG oslo_vmware.api [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130193, 'name': ReconfigVM_Task, 'duration_secs': 0.230882} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.590623] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246941', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'name': 'volume-87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0745dd19-6b08-447a-8333-047aeaa1fa83', 'attached_at': '', 'detached_at': '', 'volume_id': '87ec6079-a5ef-48d8-bfef-9127fcf399c1', 'serial': '87ec6079-a5ef-48d8-bfef-9127fcf399c1'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 965.894363] env[61991]: DEBUG nova.objects.instance [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'flavor' on Instance uuid 5364f93c-ca53-480b-8397-c331dfdeed3f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.983964] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.144843] env[61991]: DEBUG nova.objects.instance [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'flavor' on Instance uuid 0745dd19-6b08-447a-8333-047aeaa1fa83 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.356205] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.402s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.363151] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.389s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.364101] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.368495] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.402s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.368495] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.370180] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.184s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.402743] env[61991]: DEBUG oslo_concurrency.lockutils [None req-27fddfbc-bbbe-420b-9630-c66b96d472f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.791s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.410519] env[61991]: INFO nova.scheduler.client.report [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted allocations for instance b3a2ea40-5c4c-4e7d-95b1-38a18c429cec [ 966.410519] env[61991]: INFO nova.scheduler.client.report [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Deleted allocations for instance 69a272a7-8da6-4fb2-95be-85cfb7d70d44 [ 966.485743] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.885101] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "7ad63ade-cff7-4a32-b567-783db726de55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.885532] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "7ad63ade-cff7-4a32-b567-783db726de55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.886977] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "7ad63ade-cff7-4a32-b567-783db726de55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.886977] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "7ad63ade-cff7-4a32-b567-783db726de55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.886977] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "7ad63ade-cff7-4a32-b567-783db726de55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.893094] env[61991]: INFO nova.compute.manager [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Terminating instance [ 966.897202] env[61991]: DEBUG nova.compute.manager [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.897202] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.897561] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e827209b-4ec9-4439-8c5f-7c9ab3bf179d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.907958] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.910976] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed3dc1ad-a8ed-4748-b2bf-c33d4ef69415 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.926969] env[61991]: DEBUG oslo_concurrency.lockutils [None req-31096013-b609-4dc9-b45c-fdb02571e88a tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "b3a2ea40-5c4c-4e7d-95b1-38a18c429cec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.921s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.928865] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 966.928865] env[61991]: value = "task-1130196" [ 966.928865] env[61991]: _type = "Task" [ 966.928865] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.929053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fa8dd6d-ec97-4ebd-ab7f-cd575dac8510 tempest-ListServerFiltersTestJSON-1411073982 tempest-ListServerFiltersTestJSON-1411073982-project-member] Lock "69a272a7-8da6-4fb2-95be-85cfb7d70d44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.613s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.946163] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130196, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.985408] env[61991]: INFO nova.scheduler.client.report [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocation for migration f5a2581b-5e86-49df-9a62-2b362ef342c1 [ 966.995404] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.012630] env[61991]: INFO nova.compute.manager [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Rescuing [ 967.012759] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.012880] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.013069] env[61991]: DEBUG nova.network.neutron [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.144651] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d797689a-7287-4e12-bf6a-437913a2d89d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.151022] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a480e184-4074-4497-b7a9-8dd773a787bc tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.301s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.154948] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1530fe1-1969-4dfd-ab88-d0081f86d231 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.189922] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c48afa8-a5eb-402e-aec8-e63201d6f6cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.200414] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b995354-d12f-442d-bf92-6214b07ad4dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.216794] env[61991]: DEBUG nova.compute.provider_tree [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.448418] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130196, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.488429] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.497792] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.565s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.720449] env[61991]: DEBUG nova.scheduler.client.report [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.924719] env[61991]: DEBUG nova.network.neutron [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.950093] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130196, 'name': PowerOffVM_Task, 'duration_secs': 0.899962} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.950457] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 967.950645] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 967.951368] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b895d45a-4296-45dd-9144-14087ffd8b3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.993678] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.004889] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.005261] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.005385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "0745dd19-6b08-447a-8333-047aeaa1fa83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.005554] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.005729] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.009696] env[61991]: INFO nova.compute.manager [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Terminating instance [ 968.012394] env[61991]: DEBUG nova.compute.manager [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.012394] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.013308] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91c1d52-827d-4e3a-bdcb-6f13d269e821 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.025615] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.025929] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c56c6166-2e92-4f14-8ff8-9e5f7bf3974d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.037014] env[61991]: DEBUG oslo_vmware.api [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 968.037014] env[61991]: value = "task-1130198" [ 968.037014] env[61991]: _type = "Task" [ 968.037014] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.051555] env[61991]: DEBUG oslo_vmware.api [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.347909] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.348255] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.348457] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleting the datastore file [datastore1] 7ad63ade-cff7-4a32-b567-783db726de55 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.348801] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a2c6d60-2672-4cdd-bd53-0319f821cc63 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.358573] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 968.358573] env[61991]: value = "task-1130199" [ 968.358573] env[61991]: _type = "Task" [ 968.358573] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.369441] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130199, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.429498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.491108] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130195, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.004324} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.491469] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad/OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad.vmdk to [datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk. [ 968.491673] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Cleaning up location [datastore1] OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 968.491844] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_99a6927f-e047-483b-a387-8858ff6adaad {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.492111] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c14b7314-c4d4-47fe-a699-53fce6cb6265 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.501583] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 968.501583] env[61991]: value = "task-1130200" [ 968.501583] env[61991]: _type = "Task" [ 968.501583] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.510851] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.549460] env[61991]: DEBUG oslo_vmware.api [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130198, 'name': PowerOffVM_Task, 'duration_secs': 0.290495} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.549747] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.549918] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.550195] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9c137db-90ba-45c5-867e-4ad3ab6a1166 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.622167] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.622504] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.622703] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleting the datastore file [datastore1] 0745dd19-6b08-447a-8333-047aeaa1fa83 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.622967] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11920c8d-b939-41f0-bc8d-c7aa94b5a61d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.631377] env[61991]: DEBUG oslo_vmware.api [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 968.631377] env[61991]: value = "task-1130202" [ 968.631377] env[61991]: _type = "Task" [ 968.631377] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.640725] env[61991]: DEBUG oslo_vmware.api [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.733849] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.364s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.737028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.584s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.737336] env[61991]: DEBUG nova.objects.instance [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lazy-loading 'resources' on Instance uuid 9d1ac179-3d7e-4f0b-a80d-033eba447cab {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.870450] env[61991]: DEBUG oslo_vmware.api [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130199, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243762} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.870732] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.871114] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.871114] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.871341] env[61991]: INFO nova.compute.manager [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Took 1.97 seconds to destroy the instance on the hypervisor. [ 968.871711] env[61991]: DEBUG oslo.service.loopingcall [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.871801] env[61991]: DEBUG nova.compute.manager [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.871905] env[61991]: DEBUG nova.network.neutron [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 968.970266] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.970771] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.970771] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.970921] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.971076] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.972816] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.972816] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6c0f76b-6a53-494a-9b2c-76db15b5560a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.975220] env[61991]: INFO nova.compute.manager [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Terminating instance [ 968.977838] env[61991]: DEBUG nova.compute.manager [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.977838] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.978501] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d579109a-258c-4f44-823e-6ff1ddd60089 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.983675] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 968.983675] env[61991]: value = "task-1130203" [ 968.983675] env[61991]: _type = "Task" [ 968.983675] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.990051] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.990800] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c32e1397-5a9d-4f39-8b3a-034ff6086d33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.998551] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130203, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.000469] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 969.000469] env[61991]: value = "task-1130204" [ 969.000469] env[61991]: _type = "Task" [ 969.000469] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.022248] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.022784] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.055103} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.023213] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.023756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.023911] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk to [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.025012] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-113d7844-433c-4def-8476-248fea1dda47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.034814] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 969.034814] env[61991]: value = "task-1130205" [ 969.034814] env[61991]: _type = "Task" [ 969.034814] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.047280] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.145810] env[61991]: DEBUG oslo_vmware.api [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166417} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.146760] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.147171] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.149055] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.149055] env[61991]: INFO nova.compute.manager [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Took 1.14 seconds to destroy the instance on the hypervisor. [ 969.149055] env[61991]: DEBUG oslo.service.loopingcall [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.149055] env[61991]: DEBUG nova.compute.manager [-] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.149055] env[61991]: DEBUG nova.network.neutron [-] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 969.316277] env[61991]: INFO nova.scheduler.client.report [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocation for migration ac7f2f11-e6d1-4f68-81ba-58a9c7b870d1 [ 969.440397] env[61991]: DEBUG nova.compute.manager [req-8f2c3843-7184-44dc-8042-2bfd3cb1c46d req-8da5516f-3bc1-4d6a-bb79-2e55493ae548 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Received event network-vif-deleted-f039d7a6-60dd-49c4-95bc-ba71796b1aba {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.440599] env[61991]: INFO nova.compute.manager [req-8f2c3843-7184-44dc-8042-2bfd3cb1c46d req-8da5516f-3bc1-4d6a-bb79-2e55493ae548 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Neutron deleted interface f039d7a6-60dd-49c4-95bc-ba71796b1aba; detaching it from the instance and deleting it from the info cache [ 969.440774] env[61991]: DEBUG nova.network.neutron [req-8f2c3843-7184-44dc-8042-2bfd3cb1c46d req-8da5516f-3bc1-4d6a-bb79-2e55493ae548 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.472711] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db762a7-ec11-4418-bd40-4926f72d1380 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.482750] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8003383-1ca1-47eb-9957-8180736eec5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.535305] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d712242-be3c-442a-abc9-9c672d2856fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.538604] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130203, 'name': PowerOffVM_Task, 'duration_secs': 0.372679} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.538971] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.544992] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb61653-ad10-485b-b6a0-e0ca19c51ca6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.552648] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130204, 'name': PowerOffVM_Task, 'duration_secs': 0.258931} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.554546] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd1ad9f-a316-496e-8a68-c1413d4201a2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.558678] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.558862] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.578663] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2dc09e47-94d1-42d2-80c5-3164e2acc5af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.580533] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.583413] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2d0963-2115-45ae-a434-090e4f493b8f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.598907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "892323a3-8661-47da-a52a-d7e0557ea300" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.599193] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "892323a3-8661-47da-a52a-d7e0557ea300" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.600616] env[61991]: DEBUG nova.compute.provider_tree [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.653323] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.653952] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80793bff-5d66-4842-aab2-b487c54a3ba5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.664318] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 969.664318] env[61991]: value = "task-1130207" [ 969.664318] env[61991]: _type = "Task" [ 969.664318] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.677402] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 969.677710] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.678082] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.678154] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.678479] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.678821] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20927732-25a7-42f4-b1ef-10b70883b110 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.683677] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.683922] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.684119] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleting the datastore file [datastore2] 3b647920-4a69-4d1d-adb1-6fbf716e5514 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.684427] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0272be27-5dac-4da0-b681-0fb1d6c26184 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.686576] env[61991]: DEBUG nova.network.neutron [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.693611] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.693611] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 969.695716] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec79aa7-e9e5-4f59-a409-a458d70052cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.698974] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for the task: (returnval){ [ 969.698974] env[61991]: value = "task-1130208" [ 969.698974] env[61991]: _type = "Task" [ 969.698974] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.706738] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 969.706738] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523fa4a9-0ec4-6ddf-3c02-088350f6d94c" [ 969.706738] env[61991]: _type = "Task" [ 969.706738] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.710335] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130208, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.722026] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523fa4a9-0ec4-6ddf-3c02-088350f6d94c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.821833] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7228c864-2c9d-40e7-b9dd-685ab26d6f3f tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.558s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.945321] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45a53508-10b9-456f-892e-bc1f95b8e98c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.961054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a13c4b-d9e1-47fa-8d8c-b668550181ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.002431] env[61991]: DEBUG nova.compute.manager [req-8f2c3843-7184-44dc-8042-2bfd3cb1c46d req-8da5516f-3bc1-4d6a-bb79-2e55493ae548 service nova] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Detach interface failed, port_id=f039d7a6-60dd-49c4-95bc-ba71796b1aba, reason: Instance 7ad63ade-cff7-4a32-b567-783db726de55 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.016700] env[61991]: DEBUG nova.compute.manager [req-db6d969f-72b7-4d8a-820c-8fe4671a7b01 req-448e4200-104c-4b0b-a496-787981f0053f service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Received event network-vif-deleted-02083cbb-c0d3-414b-9882-6260c77dfd0c {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.016923] env[61991]: INFO nova.compute.manager [req-db6d969f-72b7-4d8a-820c-8fe4671a7b01 req-448e4200-104c-4b0b-a496-787981f0053f service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Neutron deleted interface 02083cbb-c0d3-414b-9882-6260c77dfd0c; detaching it from the instance and deleting it from the info cache [ 970.017175] env[61991]: DEBUG nova.network.neutron [req-db6d969f-72b7-4d8a-820c-8fe4671a7b01 req-448e4200-104c-4b0b-a496-787981f0053f service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.055940] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.103228] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.108027] env[61991]: DEBUG nova.scheduler.client.report [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.190776] env[61991]: INFO nova.compute.manager [-] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Took 1.32 seconds to deallocate network for instance. [ 970.231805] env[61991]: DEBUG oslo_vmware.api [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Task: {'id': task-1130208, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318019} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.238434] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.239069] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.239403] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.239711] env[61991]: INFO nova.compute.manager [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Took 1.26 seconds to destroy the instance on the hypervisor. [ 970.240514] env[61991]: DEBUG oslo.service.loopingcall [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.240894] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]523fa4a9-0ec4-6ddf-3c02-088350f6d94c, 'name': SearchDatastore_Task, 'duration_secs': 0.025122} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.241328] env[61991]: DEBUG nova.compute.manager [-] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.241544] env[61991]: DEBUG nova.network.neutron [-] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 970.245596] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c437ccf4-ee42-4e1c-a9cf-b9f3fff6175c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.258861] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 970.258861] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529227fd-5be2-47d9-2fd9-ccff0a08e632" [ 970.258861] env[61991]: _type = "Task" [ 970.258861] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.272352] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529227fd-5be2-47d9-2fd9-ccff0a08e632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.401992] env[61991]: DEBUG nova.network.neutron [-] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.522102] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-870df175-79de-40e9-a6d8-6bb79b321080 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.535268] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abf4aad-5438-4d64-9d1c-14a2f038b561 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.559655] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.572313] env[61991]: DEBUG nova.compute.manager [req-db6d969f-72b7-4d8a-820c-8fe4671a7b01 req-448e4200-104c-4b0b-a496-787981f0053f service nova] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Detach interface failed, port_id=02083cbb-c0d3-414b-9882-6260c77dfd0c, reason: Instance 0745dd19-6b08-447a-8333-047aeaa1fa83 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.615994] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.621982] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.978s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.622447] env[61991]: DEBUG nova.objects.instance [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61991) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 970.646800] env[61991]: INFO nova.scheduler.client.report [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Deleted allocations for instance 9d1ac179-3d7e-4f0b-a80d-033eba447cab [ 970.655055] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.717189] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.774211] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529227fd-5be2-47d9-2fd9-ccff0a08e632, 'name': SearchDatastore_Task, 'duration_secs': 0.016862} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.774622] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.774980] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. {{(pid=61991) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 970.775729] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd0aa459-c1d9-441d-9634-d9aa13c6e12e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.787223] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 970.787223] env[61991]: value = "task-1130209" [ 970.787223] env[61991]: _type = "Task" [ 970.787223] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.801049] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130209, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.889563] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.889891] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.907295] env[61991]: INFO nova.compute.manager [-] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Took 1.76 seconds to deallocate network for instance. [ 971.060752] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.158228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-61a68cf4-c8f7-4241-ab94-42458d3cbcbc tempest-AttachInterfacesTestJSON-1131306033 tempest-AttachInterfacesTestJSON-1131306033-project-member] Lock "9d1ac179-3d7e-4f0b-a80d-033eba447cab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.954s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.183071] env[61991]: DEBUG nova.network.neutron [-] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.301358] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130209, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.393037] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.416690] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.561152] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.635078] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0ba1682a-018c-4b9d-b867-c28e0b408903 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.636493] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.983s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.638973] env[61991]: INFO nova.compute.claims [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.689577] env[61991]: INFO nova.compute.manager [-] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Took 1.45 seconds to deallocate network for instance. [ 971.713697] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "68268135-4cf4-4732-9104-1720bbb21acf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.713887] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.714152] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "68268135-4cf4-4732-9104-1720bbb21acf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.714331] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.714510] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.717299] env[61991]: INFO nova.compute.manager [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Terminating instance [ 971.719491] env[61991]: DEBUG nova.compute.manager [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.719581] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.720458] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2dfcf74-a738-4bcd-a412-5391107a0c7d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.735605] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.737067] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02923189-1508-44eb-be56-18a8ad2a4174 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.749753] env[61991]: DEBUG oslo_vmware.api [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 971.749753] env[61991]: value = "task-1130210" [ 971.749753] env[61991]: _type = "Task" [ 971.749753] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.763167] env[61991]: DEBUG oslo_vmware.api [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.804481] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130209, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747478} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.804658] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk. [ 971.805601] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af31f96d-6360-42a5-881f-54c498608181 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.839985] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.840347] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f77d17ca-c03e-4658-96a1-3d5f84c25617 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.864694] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 971.864694] env[61991]: value = "task-1130211" [ 971.864694] env[61991]: _type = "Task" [ 971.864694] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.878941] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130211, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.919427] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.060594] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130205, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.819767} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.060964] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5cf3221b-0c05-4b11-a46c-192b39b794ad/5cf3221b-0c05-4b11-a46c-192b39b794ad.vmdk to [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 972.062014] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc87b212-594c-453e-a5f2-d0017fb9f880 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.089019] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.089392] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38da5758-16c9-47ad-94a9-4ea1ce14af64 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.110858] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 972.110858] env[61991]: value = "task-1130212" [ 972.110858] env[61991]: _type = "Task" [ 972.110858] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.120115] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130212, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.198502] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.262804] env[61991]: DEBUG oslo_vmware.api [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130210, 'name': PowerOffVM_Task, 'duration_secs': 0.326622} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.263570] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.263932] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.264364] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dfc5f14d-0a5c-40b4-a893-d68659188ea5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.293323] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.293323] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.319938] env[61991]: DEBUG nova.compute.manager [req-ce62dd63-90d6-4805-aed8-22b27b37c835 req-ac7405ad-e0ef-474d-bcef-368f3db93a67 service nova] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Received event network-vif-deleted-2dc3450f-efd8-4200-a182-988386f68d00 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.347423] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.347423] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.347423] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore2] 68268135-4cf4-4732-9104-1720bbb21acf {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.348036] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3120b584-3b2e-423c-9265-80a0d79ad841 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.359128] env[61991]: DEBUG oslo_vmware.api [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 972.359128] env[61991]: value = "task-1130214" [ 972.359128] env[61991]: _type = "Task" [ 972.359128] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.368339] env[61991]: DEBUG oslo_vmware.api [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.381220] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130211, 'name': ReconfigVM_Task, 'duration_secs': 0.409917} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.381662] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f/254d700f-2f5a-49a3-8762-cec07162124a-rescue.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.382595] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b1a29a-6f52-40c1-a35c-42124c3e91d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.412146] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7007d33e-1d8c-400c-b653-86988614862d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.428703] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 972.428703] env[61991]: value = "task-1130215" [ 972.428703] env[61991]: _type = "Task" [ 972.428703] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.439664] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.621450] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130212, 'name': ReconfigVM_Task, 'duration_secs': 0.351015} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.621752] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4/56188232-bed0-4a4b-a4bc-01edbb85cbe4.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.622407] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97905d61-d83e-42dc-86c0-a850b4bce49e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.630028] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 972.630028] env[61991]: value = "task-1130216" [ 972.630028] env[61991]: _type = "Task" [ 972.630028] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.639870] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130216, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.795500] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.863207] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66afc70-8fa7-46ab-b46c-21b02e51c288 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.875534] env[61991]: DEBUG oslo_vmware.api [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224206} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.875534] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.875534] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.875534] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.875791] env[61991]: INFO nova.compute.manager [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Took 1.16 seconds to destroy the instance on the hypervisor. [ 972.875851] env[61991]: DEBUG oslo.service.loopingcall [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.876890] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbb0dba-8685-4230-bf1b-60dc77b81c18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.880046] env[61991]: DEBUG nova.compute.manager [-] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.880154] env[61991]: DEBUG nova.network.neutron [-] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.910767] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb985063-3c0e-4af5-a7d7-a7fcfe098670 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.919170] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea5e738-9069-43c6-9304-a648c630f6e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.934724] env[61991]: DEBUG nova.compute.provider_tree [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.947731] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130215, 'name': ReconfigVM_Task, 'duration_secs': 0.213195} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.947997] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.948264] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c81e66e9-b704-4ce1-85a0-f4d5bf8cf680 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.956038] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 972.956038] env[61991]: value = "task-1130217" [ 972.956038] env[61991]: _type = "Task" [ 972.956038] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.964183] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.140553] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130216, 'name': Rename_Task, 'duration_secs': 0.202099} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.140779] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 973.141050] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-900390ac-6926-4b7d-9f4f-50097e09695d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.150107] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 973.150107] env[61991]: value = "task-1130218" [ 973.150107] env[61991]: _type = "Task" [ 973.150107] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.160120] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130218, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.323859] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.466669] env[61991]: DEBUG oslo_vmware.api [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130217, 'name': PowerOnVM_Task, 'duration_secs': 0.442189} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.467787] env[61991]: ERROR nova.scheduler.client.report [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [req-c4e4cdd7-5339-4c7d-8a9a-fd6f04af941b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c4e4cdd7-5339-4c7d-8a9a-fd6f04af941b"}]} [ 973.468185] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.472384] env[61991]: DEBUG nova.compute.manager [None req-0925a8bb-e4c7-4716-8be3-f4f17a1d97ed tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.473852] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8fa14a-e3b9-40ba-8cb9-f31bd6072677 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.495541] env[61991]: DEBUG nova.scheduler.client.report [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 973.517741] env[61991]: DEBUG nova.scheduler.client.report [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 973.517997] env[61991]: DEBUG nova.compute.provider_tree [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.537810] env[61991]: DEBUG nova.scheduler.client.report [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 973.558294] env[61991]: DEBUG nova.scheduler.client.report [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 973.662158] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130218, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.742310] env[61991]: DEBUG nova.network.neutron [-] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.774049] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fa560c-2cd3-4f6b-b834-b36ae5cb55fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.784946] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1ec97e-d870-4fc7-b4e0-fd2fe3c5c8be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.825408] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a50430f-4a35-4229-bb93-4fc8a761f5c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.833827] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d9c3fe-11fe-4ed0-827b-80e1f6cc067d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.847453] env[61991]: DEBUG nova.compute.provider_tree [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.162230] env[61991]: DEBUG oslo_vmware.api [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130218, 'name': PowerOnVM_Task, 'duration_secs': 0.6715} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.162522] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.245319] env[61991]: INFO nova.compute.manager [-] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Took 1.36 seconds to deallocate network for instance. [ 974.267226] env[61991]: DEBUG nova.compute.manager [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.268166] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5589bc-e3d8-4ee7-96b4-7a45f8379619 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.346182] env[61991]: DEBUG nova.compute.manager [req-7ed0fb0e-a141-42dc-922e-8d084233d821 req-74a84ed1-f016-4daa-88c8-8941688b047b service nova] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Received event network-vif-deleted-a511534a-b3ea-4087-81d0-c1b3277ecb0f {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.380743] env[61991]: DEBUG nova.scheduler.client.report [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 137 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 974.381020] env[61991]: DEBUG nova.compute.provider_tree [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 137 to 138 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 974.381262] env[61991]: DEBUG nova.compute.provider_tree [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.752605] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.786565] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ab6e6131-d92d-4ada-aeec-a1538c51a0fd tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.160s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.887062] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.250s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.887182] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.890795] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.174s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.891081] env[61991]: DEBUG nova.objects.instance [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'resources' on Instance uuid 7ad63ade-cff7-4a32-b567-783db726de55 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.179538] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "679d3629-63e0-452f-9cee-1da4b9b843de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.179810] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "679d3629-63e0-452f-9cee-1da4b9b843de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.394046] env[61991]: DEBUG nova.compute.utils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.398270] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.398381] env[61991]: DEBUG nova.network.neutron [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 975.430823] env[61991]: INFO nova.compute.manager [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Unrescuing [ 975.431095] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.431259] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.431430] env[61991]: DEBUG nova.network.neutron [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 975.447184] env[61991]: DEBUG nova.policy [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60ce5423a6084c50b41a46ec32019589', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c6da5920890493ca72094bcf15a73d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.599009] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1d8ea3-074f-4d07-9e8d-417b851e96fa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.606803] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8b6754-8fe5-446d-b91d-0a954e82c451 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.639902] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f50493-02b9-47ac-ac6b-3fc686b54769 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.647643] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c90835-e529-4c92-ac68-24a031f239c5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.662399] env[61991]: DEBUG nova.compute.provider_tree [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.682590] env[61991]: DEBUG nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.759675] env[61991]: DEBUG nova.network.neutron [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Successfully created port: 4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.900255] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 976.044953] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4accfa11-e24b-469e-8197-9bd25b9b1ce0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.053172] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Suspending the VM {{(pid=61991) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 976.053487] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-fae95abe-a53d-47e2-b110-552f24b1f3f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.061122] env[61991]: DEBUG oslo_vmware.api [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 976.061122] env[61991]: value = "task-1130219" [ 976.061122] env[61991]: _type = "Task" [ 976.061122] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.069639] env[61991]: DEBUG oslo_vmware.api [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130219, 'name': SuspendVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.165964] env[61991]: DEBUG nova.scheduler.client.report [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.205705] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.248243] env[61991]: DEBUG nova.network.neutron [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.472228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.472490] env[61991]: DEBUG oslo_concurrency.lockutils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.572571] env[61991]: DEBUG oslo_vmware.api [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130219, 'name': SuspendVM_Task} progress is 62%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.674192] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.783s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.676446] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.261s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.676876] env[61991]: DEBUG nova.objects.instance [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'resources' on Instance uuid 0745dd19-6b08-447a-8333-047aeaa1fa83 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.694026] env[61991]: INFO nova.scheduler.client.report [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted allocations for instance 7ad63ade-cff7-4a32-b567-783db726de55 [ 976.751124] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.751871] env[61991]: DEBUG nova.objects.instance [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'flavor' on Instance uuid 5364f93c-ca53-480b-8397-c331dfdeed3f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.914908] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.940795] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.941072] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.941240] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.941431] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.941613] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.941724] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.941933] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.942108] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.942281] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.942448] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.942630] env[61991]: DEBUG nova.virt.hardware [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.943525] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff1c520-203c-494d-a04e-928b9b728af3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.951344] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f148ec-6a02-4266-8c65-298065dfceb1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.975653] env[61991]: DEBUG nova.compute.utils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.071538] env[61991]: DEBUG oslo_vmware.api [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130219, 'name': SuspendVM_Task, 'duration_secs': 0.680711} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.071820] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Suspended the VM {{(pid=61991) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 977.072041] env[61991]: DEBUG nova.compute.manager [None req-d9c0a0f8-379f-48b8-9af6-ecafdb0a036c tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.072749] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15578828-0dc4-462f-aa68-65df014451c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.204555] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3e1189bb-6a1b-49c1-9d75-168c92d98f65 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "7ad63ade-cff7-4a32-b567-783db726de55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.319s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.256532] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6452f49d-d596-4780-88c7-40f3434c62d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.280820] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 977.283439] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a8a7c6c-7517-4393-805e-09a54b637ba1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.290491] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 977.290491] env[61991]: value = "task-1130220" [ 977.290491] env[61991]: _type = "Task" [ 977.290491] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.299939] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.360421] env[61991]: DEBUG nova.network.neutron [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Successfully updated port: 4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.405277] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ded013-b7f2-4d0c-820d-823c87f6d8b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.413297] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006d9848-3eaa-4a82-a3e9-1d3a6251652a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.443292] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9b5778-9289-41ad-9cf7-7ed3ac1ca7e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.450401] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b2e9a5-ed14-4f58-b6f7-b93322be7194 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.464426] env[61991]: DEBUG nova.compute.provider_tree [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.478626] env[61991]: DEBUG oslo_concurrency.lockutils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.606822] env[61991]: DEBUG nova.compute.manager [req-0a800ef0-8226-4def-9237-f24f8e1f42ae req-3634f0c3-6d1f-442a-b5ed-a21de6c27346 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Received event network-vif-plugged-4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.607067] env[61991]: DEBUG oslo_concurrency.lockutils [req-0a800ef0-8226-4def-9237-f24f8e1f42ae req-3634f0c3-6d1f-442a-b5ed-a21de6c27346 service nova] Acquiring lock "892323a3-8661-47da-a52a-d7e0557ea300-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.607289] env[61991]: DEBUG oslo_concurrency.lockutils [req-0a800ef0-8226-4def-9237-f24f8e1f42ae req-3634f0c3-6d1f-442a-b5ed-a21de6c27346 service nova] Lock "892323a3-8661-47da-a52a-d7e0557ea300-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.607464] env[61991]: DEBUG oslo_concurrency.lockutils [req-0a800ef0-8226-4def-9237-f24f8e1f42ae req-3634f0c3-6d1f-442a-b5ed-a21de6c27346 service nova] Lock "892323a3-8661-47da-a52a-d7e0557ea300-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.607675] env[61991]: DEBUG nova.compute.manager [req-0a800ef0-8226-4def-9237-f24f8e1f42ae req-3634f0c3-6d1f-442a-b5ed-a21de6c27346 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] No waiting events found dispatching network-vif-plugged-4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.607796] env[61991]: WARNING nova.compute.manager [req-0a800ef0-8226-4def-9237-f24f8e1f42ae req-3634f0c3-6d1f-442a-b5ed-a21de6c27346 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Received unexpected event network-vif-plugged-4e71b591-340e-4029-8776-b78a3dc981ce for instance with vm_state building and task_state spawning. [ 977.801594] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130220, 'name': PowerOffVM_Task, 'duration_secs': 0.229726} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.801880] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.807028] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfiguring VM instance instance-0000005c to detach disk 2002 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 977.807333] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60635bfb-729a-4b07-ae2a-32fee2e4f355 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.824363] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 977.824363] env[61991]: value = "task-1130221" [ 977.824363] env[61991]: _type = "Task" [ 977.824363] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.832136] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130221, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.862424] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.862587] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquired lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.862744] env[61991]: DEBUG nova.network.neutron [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.967943] env[61991]: DEBUG nova.scheduler.client.report [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.334436] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130221, 'name': ReconfigVM_Task, 'duration_secs': 0.232322} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.334714] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfigured VM instance instance-0000005c to detach disk 2002 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 978.334907] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.335181] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a61b898-402c-420a-98e4-49dab8f40c16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.341717] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 978.341717] env[61991]: value = "task-1130222" [ 978.341717] env[61991]: _type = "Task" [ 978.341717] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.348742] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130222, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.395023] env[61991]: DEBUG nova.network.neutron [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.473340] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.797s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.475602] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.556s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.477068] env[61991]: INFO nova.compute.claims [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.490906] env[61991]: INFO nova.scheduler.client.report [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted allocations for instance 0745dd19-6b08-447a-8333-047aeaa1fa83 [ 978.550051] env[61991]: DEBUG oslo_concurrency.lockutils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.550181] env[61991]: DEBUG oslo_concurrency.lockutils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.550408] env[61991]: INFO nova.compute.manager [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Attaching volume 1b159140-cae6-4f5b-9849-8a3da257f4bb to /dev/sdb [ 978.577141] env[61991]: DEBUG nova.network.neutron [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updating instance_info_cache with network_info: [{"id": "4e71b591-340e-4029-8776-b78a3dc981ce", "address": "fa:16:3e:b5:95:c9", "network": {"id": "e869a986-5ddf-428c-893a-1ce59b6537ec", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-132331547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6da5920890493ca72094bcf15a73d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e71b591-34", "ovs_interfaceid": "4e71b591-340e-4029-8776-b78a3dc981ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.584681] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d9945a-fd93-4cb9-bb72-26e3b51590ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.591991] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb67e08-3c22-4be8-bba0-8dd881b54662 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.606175] env[61991]: DEBUG nova.virt.block_device [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating existing volume attachment record: 8dce2689-5cfd-4b64-99ba-3ce75d0c7e2b {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 978.691215] env[61991]: INFO nova.compute.manager [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Resuming [ 978.692314] env[61991]: DEBUG nova.objects.instance [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'flavor' on Instance uuid 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.852641] env[61991]: DEBUG oslo_vmware.api [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130222, 'name': PowerOnVM_Task, 'duration_secs': 0.36585} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.852938] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.853220] env[61991]: DEBUG nova.compute.manager [None req-c45a67fc-6c79-45a6-be4e-88f64ceb62a9 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.853983] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e0f655-60d7-4dc7-a639-f5ed70d85e9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.973627] env[61991]: DEBUG nova.compute.manager [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 978.998632] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f80a7869-807d-471f-9494-e385fc631bf3 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "0745dd19-6b08-447a-8333-047aeaa1fa83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.993s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.079429] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Releasing lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.079763] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Instance network_info: |[{"id": "4e71b591-340e-4029-8776-b78a3dc981ce", "address": "fa:16:3e:b5:95:c9", "network": {"id": "e869a986-5ddf-428c-893a-1ce59b6537ec", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-132331547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6da5920890493ca72094bcf15a73d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e71b591-34", "ovs_interfaceid": "4e71b591-340e-4029-8776-b78a3dc981ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.080250] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:95:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6edb8eae-1113-49d0-84f7-9fd9f82b26fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e71b591-340e-4029-8776-b78a3dc981ce', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.089026] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Creating folder: Project (9c6da5920890493ca72094bcf15a73d3). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 979.089361] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3780c763-c69c-4a74-b0d5-f923cce67ce7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.099790] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Created folder: Project (9c6da5920890493ca72094bcf15a73d3) in parent group-v246753. [ 979.100062] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Creating folder: Instances. Parent ref: group-v246949. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 979.100331] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cc9f285-69a1-467a-8f88-0b47768bfcf8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.108826] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Created folder: Instances in parent group-v246949. [ 979.109121] env[61991]: DEBUG oslo.service.loopingcall [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.109378] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.109632] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c803d26-deec-402c-80e9-da8da9270fac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.128594] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.128594] env[61991]: value = "task-1130228" [ 979.128594] env[61991]: _type = "Task" [ 979.128594] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.137447] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130228, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.496333] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.637822] env[61991]: DEBUG nova.compute.manager [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Received event network-changed-4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.638630] env[61991]: DEBUG nova.compute.manager [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Refreshing instance network info cache due to event network-changed-4e71b591-340e-4029-8776-b78a3dc981ce. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.638630] env[61991]: DEBUG oslo_concurrency.lockutils [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] Acquiring lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.638630] env[61991]: DEBUG oslo_concurrency.lockutils [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] Acquired lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.638630] env[61991]: DEBUG nova.network.neutron [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Refreshing network info cache for port 4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 979.646611] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130228, 'name': CreateVM_Task, 'duration_secs': 0.337152} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.647276] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.648017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.648201] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.648542] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.649010] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7015771-72e1-479d-873d-1f58c9d995f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.657485] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 979.657485] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522ddd40-2bb6-0a80-f28c-497ec30b963e" [ 979.657485] env[61991]: _type = "Task" [ 979.657485] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.663843] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d972231c-2051-4367-b1dd-5c4964c90d3b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.669846] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522ddd40-2bb6-0a80-f28c-497ec30b963e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.675348] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1025a6-edb3-4577-b431-a252105ef04f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.705895] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.706113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquired lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.706317] env[61991]: DEBUG nova.network.neutron [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 979.708185] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da795abf-a072-4af7-8700-54541a75c22d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.715710] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa165069-efe3-4429-b54f-cf1ac3e5858d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.729440] env[61991]: DEBUG nova.compute.provider_tree [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.171517] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522ddd40-2bb6-0a80-f28c-497ec30b963e, 'name': SearchDatastore_Task, 'duration_secs': 0.016083} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.171517] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.171517] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.171517] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.171517] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.171517] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.171517] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3192907-3a94-428b-bfea-bc1b964408f2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.183398] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.184693] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 980.184693] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee870c3c-c73e-4a4f-91b0-e3d4321fcb16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.189880] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 980.189880] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52028945-7419-6df2-50ea-f014a358e6a0" [ 980.189880] env[61991]: _type = "Task" [ 980.189880] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.200352] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52028945-7419-6df2-50ea-f014a358e6a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.231891] env[61991]: DEBUG nova.scheduler.client.report [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.531103] env[61991]: DEBUG nova.network.neutron [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [{"id": "c7263e46-a164-4813-be3e-faae7fc31ff2", "address": "fa:16:3e:b9:d4:42", "network": {"id": "3138206d-a8aa-411f-b6f4-da14b11c61fd", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-783678382-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a2b37cc1ade437097583e823d457ca6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7263e46-a1", "ovs_interfaceid": "c7263e46-a164-4813-be3e-faae7fc31ff2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.543223] env[61991]: DEBUG nova.network.neutron [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updated VIF entry in instance network info cache for port 4e71b591-340e-4029-8776-b78a3dc981ce. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 980.543620] env[61991]: DEBUG nova.network.neutron [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updating instance_info_cache with network_info: [{"id": "4e71b591-340e-4029-8776-b78a3dc981ce", "address": "fa:16:3e:b5:95:c9", "network": {"id": "e869a986-5ddf-428c-893a-1ce59b6537ec", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-132331547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6da5920890493ca72094bcf15a73d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e71b591-34", "ovs_interfaceid": "4e71b591-340e-4029-8776-b78a3dc981ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.700956] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52028945-7419-6df2-50ea-f014a358e6a0, 'name': SearchDatastore_Task, 'duration_secs': 0.01406} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.701438] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de330493-c60a-4966-b25d-f8704d92c09c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.708099] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 980.708099] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b2448c-e095-d306-4330-3bc4e8bd717f" [ 980.708099] env[61991]: _type = "Task" [ 980.708099] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.714439] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b2448c-e095-d306-4330-3bc4e8bd717f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.736381] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.736918] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.739550] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.542s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.740078] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.742831] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.418s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.744363] env[61991]: INFO nova.compute.claims [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.769769] env[61991]: INFO nova.scheduler.client.report [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Deleted allocations for instance 3b647920-4a69-4d1d-adb1-6fbf716e5514 [ 981.035252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Releasing lock "refresh_cache-56188232-bed0-4a4b-a4bc-01edbb85cbe4" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.036113] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f1ae21-91f0-483a-9d63-760da572e719 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.043023] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Resuming the VM {{(pid=61991) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 981.043321] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab113207-52e2-4c2e-98d4-474bc3e5d0e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.046602] env[61991]: DEBUG oslo_concurrency.lockutils [req-a3b87bd3-db23-441f-8206-d9dabe4c694d req-dfa050f3-0598-4454-b713-29f3839d71f8 service nova] Releasing lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.049075] env[61991]: DEBUG oslo_vmware.api [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 981.049075] env[61991]: value = "task-1130230" [ 981.049075] env[61991]: _type = "Task" [ 981.049075] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.057052] env[61991]: DEBUG oslo_vmware.api [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130230, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.217348] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b2448c-e095-d306-4330-3bc4e8bd717f, 'name': SearchDatastore_Task, 'duration_secs': 0.0094} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.217725] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.218072] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 892323a3-8661-47da-a52a-d7e0557ea300/892323a3-8661-47da-a52a-d7e0557ea300.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.218420] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d0799ba-2abf-442c-a37d-dc8a3fe32499 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.224741] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 981.224741] env[61991]: value = "task-1130231" [ 981.224741] env[61991]: _type = "Task" [ 981.224741] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.232323] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.248802] env[61991]: DEBUG nova.compute.utils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.252706] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.252706] env[61991]: DEBUG nova.network.neutron [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 981.280375] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3f969e53-6459-4fee-ab68-4ee01ce01fd1 tempest-DeleteServersTestJSON-1972816297 tempest-DeleteServersTestJSON-1972816297-project-member] Lock "3b647920-4a69-4d1d-adb1-6fbf716e5514" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.310s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.317854] env[61991]: DEBUG nova.policy [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b5d90a913fc41e282e050bb9af628d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adbd4ea9b0744f28a0d4a46b4a04d683', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.403477] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.403905] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.561242] env[61991]: DEBUG oslo_vmware.api [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130230, 'name': PowerOnVM_Task, 'duration_secs': 0.488612} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.561453] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Resumed the VM {{(pid=61991) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 981.561679] env[61991]: DEBUG nova.compute.manager [None req-ca5c4fea-1eef-4e23-81fb-44df0d68328b tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 981.562577] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e401d34-1b6a-4305-91f9-a703617507ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.735200] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130231, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467386} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.735569] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 892323a3-8661-47da-a52a-d7e0557ea300/892323a3-8661-47da-a52a-d7e0557ea300.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.735791] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.736082] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be428137-4eff-451e-a202-0e1fdb3d9a25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.742849] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 981.742849] env[61991]: value = "task-1130232" [ 981.742849] env[61991]: _type = "Task" [ 981.742849] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.751626] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.756308] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.855853] env[61991]: DEBUG nova.compute.manager [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.856131] env[61991]: DEBUG nova.compute.manager [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing instance network info cache due to event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.856358] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.856514] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.856638] env[61991]: DEBUG nova.network.neutron [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.907158] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 981.959317] env[61991]: DEBUG nova.network.neutron [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Successfully created port: 96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.999592] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e8def0-e3b9-4062-bd3a-79e9bc69d652 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.007911] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ee4ff2-9015-44f0-bcef-36dab1ae9031 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.038406] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21abd1f-d7f0-45f7-bb4d-f2334761896f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.046360] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85e9de3-906f-4ac1-8dd7-b0bea35a6b7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.059928] env[61991]: DEBUG nova.compute.provider_tree [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.254652] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079213} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.254976] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.255788] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce15bc07-5911-40bc-a928-cb554b766d30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.281366] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 892323a3-8661-47da-a52a-d7e0557ea300/892323a3-8661-47da-a52a-d7e0557ea300.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.281882] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e875653b-0e54-43ae-a296-56e6c824aecd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.302340] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 982.302340] env[61991]: value = "task-1130233" [ 982.302340] env[61991]: _type = "Task" [ 982.302340] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.311416] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130233, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.438194] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.563102] env[61991]: DEBUG nova.scheduler.client.report [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.783608] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.790415] env[61991]: DEBUG nova.network.neutron [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updated VIF entry in instance network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.790779] env[61991]: DEBUG nova.network.neutron [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.817616] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130233, 'name': ReconfigVM_Task, 'duration_secs': 0.352526} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.819875] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.820142] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.820310] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.820608] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.820716] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.820824] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.821078] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.821231] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.821434] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.822027] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.822027] env[61991]: DEBUG nova.virt.hardware [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.822126] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 892323a3-8661-47da-a52a-d7e0557ea300/892323a3-8661-47da-a52a-d7e0557ea300.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.823437] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5586e295-9911-47ef-8b93-76d5f161f828 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.825881] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1a67d3a-ed76-4080-9ae9-d419c23fb758 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.836216] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab645aa-0b1f-4932-b45e-12fbb2ff8e20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.842887] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 982.842887] env[61991]: value = "task-1130234" [ 982.842887] env[61991]: _type = "Task" [ 982.842887] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.865591] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130234, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.070215] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.070830] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.077613] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.325s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.077824] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.079947] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.874s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.086258] env[61991]: INFO nova.compute.claims [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.115751] env[61991]: INFO nova.scheduler.client.report [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocations for instance 68268135-4cf4-4732-9104-1720bbb21acf [ 983.158714] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 983.158973] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246948', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'name': 'volume-1b159140-cae6-4f5b-9849-8a3da257f4bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1730f9b3-69ad-476b-b3ab-3e1345f2a115', 'attached_at': '', 'detached_at': '', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'serial': '1b159140-cae6-4f5b-9849-8a3da257f4bb'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 983.159877] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192a17b8-b72b-4d30-8d14-ac5b25e7d014 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.177875] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72eeb6c-d2a6-43c8-a136-cfc7d9bdd812 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.205377] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-1b159140-cae6-4f5b-9849-8a3da257f4bb/volume-1b159140-cae6-4f5b-9849-8a3da257f4bb.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.205854] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebd42058-53d2-4c87-ae23-bd8bb22f4eef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.225078] env[61991]: DEBUG oslo_vmware.api [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 983.225078] env[61991]: value = "task-1130235" [ 983.225078] env[61991]: _type = "Task" [ 983.225078] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.232796] env[61991]: DEBUG oslo_vmware.api [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130235, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.293708] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.293996] env[61991]: DEBUG nova.compute.manager [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.294212] env[61991]: DEBUG nova.compute.manager [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing instance network info cache due to event network-changed-2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.294440] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] Acquiring lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.294625] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] Acquired lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.294741] env[61991]: DEBUG nova.network.neutron [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Refreshing network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 983.354719] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130234, 'name': Rename_Task, 'duration_secs': 0.133645} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.354719] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.354719] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-393e363d-207c-40ca-8614-f690bde2a11f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.360169] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 983.360169] env[61991]: value = "task-1130236" [ 983.360169] env[61991]: _type = "Task" [ 983.360169] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.371801] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.548574] env[61991]: DEBUG nova.network.neutron [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Successfully updated port: 96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.581646] env[61991]: DEBUG nova.compute.utils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.585367] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.585367] env[61991]: DEBUG nova.network.neutron [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 983.629296] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7e288899-bc82-4529-8e8e-284306a32e03 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "68268135-4cf4-4732-9104-1720bbb21acf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.915s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.693860] env[61991]: DEBUG nova.policy [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6422dfa4c424fbaab78c3e72a0e6669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dddb51029854105bc6c9c8724181d39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.736044] env[61991]: DEBUG oslo_vmware.api [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130235, 'name': ReconfigVM_Task, 'duration_secs': 0.381428} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.736468] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-1b159140-cae6-4f5b-9849-8a3da257f4bb/volume-1b159140-cae6-4f5b-9849-8a3da257f4bb.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.741697] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-553aad53-7b54-4e77-aedf-175abe4edd7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.758685] env[61991]: DEBUG oslo_vmware.api [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 983.758685] env[61991]: value = "task-1130237" [ 983.758685] env[61991]: _type = "Task" [ 983.758685] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.771573] env[61991]: DEBUG oslo_vmware.api [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130237, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.871422] env[61991]: DEBUG oslo_vmware.api [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130236, 'name': PowerOnVM_Task, 'duration_secs': 0.438377} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.871807] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.872116] env[61991]: INFO nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Took 6.96 seconds to spawn the instance on the hypervisor. [ 983.872298] env[61991]: DEBUG nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.873552] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6377df88-ef2a-4079-afc2-0801d4e49ded {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.927936] env[61991]: DEBUG nova.compute.manager [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Received event network-vif-plugged-96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.928279] env[61991]: DEBUG oslo_concurrency.lockutils [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] Acquiring lock "100d5fc0-e9d9-4892-9f60-99d657c533ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.928602] env[61991]: DEBUG oslo_concurrency.lockutils [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.928919] env[61991]: DEBUG oslo_concurrency.lockutils [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.930305] env[61991]: DEBUG nova.compute.manager [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] No waiting events found dispatching network-vif-plugged-96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.930760] env[61991]: WARNING nova.compute.manager [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Received unexpected event network-vif-plugged-96d11c6e-409f-4baf-983b-c066592aa09b for instance with vm_state building and task_state spawning. [ 983.930830] env[61991]: DEBUG nova.compute.manager [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Received event network-changed-96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.931021] env[61991]: DEBUG nova.compute.manager [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Refreshing instance network info cache due to event network-changed-96d11c6e-409f-4baf-983b-c066592aa09b. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.933102] env[61991]: DEBUG oslo_concurrency.lockutils [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] Acquiring lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.933102] env[61991]: DEBUG oslo_concurrency.lockutils [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] Acquired lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.933102] env[61991]: DEBUG nova.network.neutron [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Refreshing network info cache for port 96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.051164] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.087836] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.101276] env[61991]: DEBUG nova.network.neutron [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Successfully created port: 15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.172083] env[61991]: DEBUG nova.network.neutron [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updated VIF entry in instance network info cache for port 2fc2099f-0ea4-4665-948a-01d944f9275a. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 984.172468] env[61991]: DEBUG nova.network.neutron [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [{"id": "2fc2099f-0ea4-4665-948a-01d944f9275a", "address": "fa:16:3e:5e:2f:5f", "network": {"id": "2f80daad-8072-4a77-96ab-bd045a92b7ff", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1611678829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c89707d8b26430c830449ab9bca4a62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc2099f-0e", "ovs_interfaceid": "2fc2099f-0ea4-4665-948a-01d944f9275a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.271278] env[61991]: DEBUG oslo_vmware.api [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130237, 'name': ReconfigVM_Task, 'duration_secs': 0.139287} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.271278] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246948', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'name': 'volume-1b159140-cae6-4f5b-9849-8a3da257f4bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1730f9b3-69ad-476b-b3ab-3e1345f2a115', 'attached_at': '', 'detached_at': '', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'serial': '1b159140-cae6-4f5b-9849-8a3da257f4bb'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 984.316552] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d8852d-2e10-41d3-89b4-35ae0c852e25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.324611] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fba48c-34b2-41dd-88ac-35a6bd1436b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.355450] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b1ce15-96ec-4a78-835e-1ffa96eaa985 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.365806] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b553000c-2714-4236-81d0-c7543a132cdd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.380050] env[61991]: DEBUG nova.compute.provider_tree [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.395857] env[61991]: INFO nova.compute.manager [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Took 13.78 seconds to build instance. [ 984.483200] env[61991]: DEBUG nova.network.neutron [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 984.557967] env[61991]: DEBUG nova.network.neutron [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.676120] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ef4767b-39dc-4af7-b326-3882693e13bc req-37817cc1-c5a1-44aa-bf3f-6a0271162677 service nova] Releasing lock "refresh_cache-5364f93c-ca53-480b-8397-c331dfdeed3f" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.885716] env[61991]: DEBUG nova.scheduler.client.report [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.905098] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8945767d-c010-421e-9fdb-8a71ec6cceef tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "892323a3-8661-47da-a52a-d7e0557ea300" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.302s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.060426] env[61991]: DEBUG oslo_concurrency.lockutils [req-e1149459-08aa-4e35-93f1-0559ac3c3a5e req-94aeb4ec-7698-40c2-a1d7-089384247e15 service nova] Releasing lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.060815] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.060979] env[61991]: DEBUG nova.network.neutron [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.096752] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.126676] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.126940] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.127116] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.127333] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.127516] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.127666] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.127875] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.128084] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.128226] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.128393] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.128568] env[61991]: DEBUG nova.virt.hardware [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.129429] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677d4b71-5d6d-45f4-a156-7780f994b204 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.137901] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147987c6-db2b-421c-bdbc-7d0f4230715b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.313063] env[61991]: DEBUG nova.objects.instance [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'flavor' on Instance uuid 1730f9b3-69ad-476b-b3ab-3e1345f2a115 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.394118] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.394118] env[61991]: DEBUG nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.397109] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.901s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.600348] env[61991]: DEBUG nova.network.neutron [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 985.818601] env[61991]: DEBUG oslo_concurrency.lockutils [None req-12f510c6-b097-49d1-8471-79d44fe171b7 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.268s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.828163] env[61991]: DEBUG nova.network.neutron [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updating instance_info_cache with network_info: [{"id": "96d11c6e-409f-4baf-983b-c066592aa09b", "address": "fa:16:3e:76:45:34", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96d11c6e-40", "ovs_interfaceid": "96d11c6e-409f-4baf-983b-c066592aa09b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.905425] env[61991]: INFO nova.compute.claims [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.911668] env[61991]: DEBUG nova.compute.utils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.914187] env[61991]: DEBUG nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Not allocating networking since 'none' was specified. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 985.971020] env[61991]: DEBUG nova.compute.manager [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Received event network-changed-4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.971020] env[61991]: DEBUG nova.compute.manager [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Refreshing instance network info cache due to event network-changed-4e71b591-340e-4029-8776-b78a3dc981ce. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.971422] env[61991]: DEBUG oslo_concurrency.lockutils [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] Acquiring lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.973144] env[61991]: DEBUG oslo_concurrency.lockutils [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] Acquired lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.973144] env[61991]: DEBUG nova.network.neutron [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Refreshing network info cache for port 4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.095739] env[61991]: DEBUG nova.network.neutron [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Successfully updated port: 15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.332248] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.332665] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Instance network_info: |[{"id": "96d11c6e-409f-4baf-983b-c066592aa09b", "address": "fa:16:3e:76:45:34", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96d11c6e-40", "ovs_interfaceid": "96d11c6e-409f-4baf-983b-c066592aa09b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.333165] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:45:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96d11c6e-409f-4baf-983b-c066592aa09b', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.345815] env[61991]: DEBUG oslo.service.loopingcall [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.345899] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.346313] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4724d24e-e67a-4b89-b45a-ed43c978e5db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.380163] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.380163] env[61991]: value = "task-1130238" [ 986.380163] env[61991]: _type = "Task" [ 986.380163] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.390645] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130238, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.416053] env[61991]: INFO nova.compute.resource_tracker [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating resource usage from migration ae50fc19-9498-4e5f-9b4f-c47537530c53 [ 986.419948] env[61991]: DEBUG nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.603502] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.603502] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.603502] env[61991]: DEBUG nova.network.neutron [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.661826] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ee16d4-9db5-4fbf-96ba-2ed212fedad8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.670145] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c90e6c-1cff-4ce4-ad4e-3b105fb8e569 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.708563] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1678f386-33c2-4b1f-b97b-5d771250bd3f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.717398] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9c0453-f06e-454e-9c21-69f5c139e126 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.731572] env[61991]: DEBUG nova.compute.provider_tree [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.886495] env[61991]: DEBUG nova.compute.manager [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 986.895090] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130238, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.911210] env[61991]: DEBUG nova.network.neutron [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updated VIF entry in instance network info cache for port 4e71b591-340e-4029-8776-b78a3dc981ce. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 986.911638] env[61991]: DEBUG nova.network.neutron [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updating instance_info_cache with network_info: [{"id": "4e71b591-340e-4029-8776-b78a3dc981ce", "address": "fa:16:3e:b5:95:c9", "network": {"id": "e869a986-5ddf-428c-893a-1ce59b6537ec", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-132331547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c6da5920890493ca72094bcf15a73d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e71b591-34", "ovs_interfaceid": "4e71b591-340e-4029-8776-b78a3dc981ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.133532] env[61991]: DEBUG nova.network.neutron [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 987.240353] env[61991]: DEBUG nova.scheduler.client.report [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.307440] env[61991]: DEBUG nova.network.neutron [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance_info_cache with network_info: [{"id": "15cc98a6-0818-4630-b737-e50e1fb71851", "address": "fa:16:3e:97:a6:51", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15cc98a6-08", "ovs_interfaceid": "15cc98a6-0818-4630-b737-e50e1fb71851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.397121] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130238, 'name': CreateVM_Task, 'duration_secs': 0.903489} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.397503] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.398233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.398506] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.398735] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.398992] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21f40175-8970-45b9-8472-958fa6316110 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.403990] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 987.403990] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526e0dd8-6161-dfae-c1b2-eb1ed6620d98" [ 987.403990] env[61991]: _type = "Task" [ 987.403990] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.413263] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526e0dd8-6161-dfae-c1b2-eb1ed6620d98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.414307] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.414912] env[61991]: DEBUG oslo_concurrency.lockutils [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] Releasing lock "refresh_cache-892323a3-8661-47da-a52a-d7e0557ea300" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.415168] env[61991]: DEBUG nova.compute.manager [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Received event network-vif-plugged-15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.415360] env[61991]: DEBUG oslo_concurrency.lockutils [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] Acquiring lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.415557] env[61991]: DEBUG oslo_concurrency.lockutils [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.415723] env[61991]: DEBUG oslo_concurrency.lockutils [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.415890] env[61991]: DEBUG nova.compute.manager [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] No waiting events found dispatching network-vif-plugged-15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.416074] env[61991]: WARNING nova.compute.manager [req-57fb443d-f01f-43b2-94cf-f035c91118f6 req-938b4562-0c51-40b0-bd74-056a15b63857 service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Received unexpected event network-vif-plugged-15cc98a6-0818-4630-b737-e50e1fb71851 for instance with vm_state building and task_state spawning. [ 987.430146] env[61991]: DEBUG nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.457597] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.457869] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.458042] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.458240] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.458612] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.458834] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.459068] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.459314] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.460084] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.460322] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.460518] env[61991]: DEBUG nova.virt.hardware [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.461386] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551f581d-22be-4425-9db3-1d2f7af1b941 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.469363] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a03000d-8f39-4bcb-8cf5-2138814684db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.481994] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Instance VIF info [] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.487514] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Creating folder: Project (cff26c94c5b743e0bd9ac35096cf9991). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.487777] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a25a8f44-b7dd-48f9-8b22-ee8a72db3f47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.497201] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Created folder: Project (cff26c94c5b743e0bd9ac35096cf9991) in parent group-v246753. [ 987.497441] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Creating folder: Instances. Parent ref: group-v246953. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.497660] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-782b27f2-601a-4bb2-9680-af9b9a2aaf06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.506338] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Created folder: Instances in parent group-v246953. [ 987.506563] env[61991]: DEBUG oslo.service.loopingcall [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.506747] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.506934] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebee934d-146c-484f-8fda-b04cb583c309 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.522533] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.522533] env[61991]: value = "task-1130241" [ 987.522533] env[61991]: _type = "Task" [ 987.522533] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.534402] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130241, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.745870] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.349s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.750112] env[61991]: INFO nova.compute.manager [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Migrating [ 987.752670] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.315s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.754227] env[61991]: INFO nova.compute.claims [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.810370] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.810732] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Instance network_info: |[{"id": "15cc98a6-0818-4630-b737-e50e1fb71851", "address": "fa:16:3e:97:a6:51", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15cc98a6-08", "ovs_interfaceid": "15cc98a6-0818-4630-b737-e50e1fb71851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.811456] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:a6:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15cc98a6-0818-4630-b737-e50e1fb71851', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.819799] env[61991]: DEBUG oslo.service.loopingcall [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.819799] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.820101] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d083b501-dab6-4fb7-86f4-a1eb9378e4e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.839248] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.839248] env[61991]: value = "task-1130242" [ 987.839248] env[61991]: _type = "Task" [ 987.839248] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.847236] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130242, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.914371] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526e0dd8-6161-dfae-c1b2-eb1ed6620d98, 'name': SearchDatastore_Task, 'duration_secs': 0.038284} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.914702] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.914931] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.915201] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.915344] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.915535] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.915824] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad9e9631-0887-4a37-a4bb-7517988e17dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.924586] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.924811] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.926023] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beed48bf-c8e9-4f41-8447-a378cd20eeb1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.931154] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 987.931154] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5249e0ed-71b3-9606-bae6-f709d87b3e6c" [ 987.931154] env[61991]: _type = "Task" [ 987.931154] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.939617] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5249e0ed-71b3-9606-bae6-f709d87b3e6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.032305] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130241, 'name': CreateVM_Task, 'duration_secs': 0.296493} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.032478] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.032914] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.033091] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.033533] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.033801] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3d3a96e-df25-45bf-b7c4-598020c9df6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.038784] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 988.038784] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5245eef5-460e-39d8-1039-e10208eb1a6c" [ 988.038784] env[61991]: _type = "Task" [ 988.038784] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.047043] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5245eef5-460e-39d8-1039-e10208eb1a6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.105023] env[61991]: DEBUG nova.compute.manager [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Received event network-changed-15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.105272] env[61991]: DEBUG nova.compute.manager [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Refreshing instance network info cache due to event network-changed-15cc98a6-0818-4630-b737-e50e1fb71851. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.105472] env[61991]: DEBUG oslo_concurrency.lockutils [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] Acquiring lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.105618] env[61991]: DEBUG oslo_concurrency.lockutils [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] Acquired lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.105777] env[61991]: DEBUG nova.network.neutron [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Refreshing network info cache for port 15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.266919] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.267153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.267344] env[61991]: DEBUG nova.network.neutron [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.285038] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.285338] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.285561] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.285749] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.285920] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.288133] env[61991]: INFO nova.compute.manager [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Terminating instance [ 988.290630] env[61991]: DEBUG nova.compute.manager [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.291670] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.292184] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda6551a-20c4-48e3-aa47-06cc9bf269ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.300062] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.301047] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-537bba6a-c834-40f4-8438-73639d04fa03 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.307119] env[61991]: DEBUG oslo_vmware.api [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 988.307119] env[61991]: value = "task-1130243" [ 988.307119] env[61991]: _type = "Task" [ 988.307119] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.316676] env[61991]: DEBUG oslo_vmware.api [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.348916] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130242, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.442577] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5249e0ed-71b3-9606-bae6-f709d87b3e6c, 'name': SearchDatastore_Task, 'duration_secs': 0.009384} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.443665] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8c67905-d375-4796-ba2c-dc37750f06f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.448988] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 988.448988] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52adfe36-0998-ecfa-9096-dc608c0f9f6e" [ 988.448988] env[61991]: _type = "Task" [ 988.448988] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.458733] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52adfe36-0998-ecfa-9096-dc608c0f9f6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.549333] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5245eef5-460e-39d8-1039-e10208eb1a6c, 'name': SearchDatastore_Task, 'duration_secs': 0.008806} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.549650] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.549886] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.550113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.816314] env[61991]: DEBUG oslo_vmware.api [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130243, 'name': PowerOffVM_Task, 'duration_secs': 0.356709} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.820107] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.820308] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.823194] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9d8e60c-cbb2-4eac-abc5-122b497a6933 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.824997] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.825241] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.826840] env[61991]: DEBUG nova.network.neutron [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updated VIF entry in instance network info cache for port 15cc98a6-0818-4630-b737-e50e1fb71851. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 988.827179] env[61991]: DEBUG nova.network.neutron [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance_info_cache with network_info: [{"id": "15cc98a6-0818-4630-b737-e50e1fb71851", "address": "fa:16:3e:97:a6:51", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15cc98a6-08", "ovs_interfaceid": "15cc98a6-0818-4630-b737-e50e1fb71851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.855207] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130242, 'name': CreateVM_Task, 'duration_secs': 0.517872} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.855453] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.856161] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.856480] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.856780] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.857071] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c993956-b5e6-4240-abdb-3805ba90fe5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.864312] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 988.864312] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b9f0f3-e262-bacf-18cb-9b39e0d44ee4" [ 988.864312] env[61991]: _type = "Task" [ 988.864312] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.873828] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b9f0f3-e262-bacf-18cb-9b39e0d44ee4, 'name': SearchDatastore_Task, 'duration_secs': 0.009051} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.876810] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.876993] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.877290] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.891495] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 988.891495] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 988.891495] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleting the datastore file [datastore1] 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.891495] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-373a604b-1faf-4dd6-ab0d-a639c8c7ca9a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.894245] env[61991]: DEBUG oslo_vmware.api [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for the task: (returnval){ [ 988.894245] env[61991]: value = "task-1130245" [ 988.894245] env[61991]: _type = "Task" [ 988.894245] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.903468] env[61991]: DEBUG oslo_vmware.api [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.959213] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52adfe36-0998-ecfa-9096-dc608c0f9f6e, 'name': SearchDatastore_Task, 'duration_secs': 0.009424} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.961747] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.962067] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 100d5fc0-e9d9-4892-9f60-99d657c533ed/100d5fc0-e9d9-4892-9f60-99d657c533ed.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 988.963379] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.963379] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.963379] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3180cc22-69c1-4ed5-a3e0-c68c2d545bbc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.964859] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7e29011-d8ca-4c26-9288-7f6f607076c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.973488] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 988.973488] env[61991]: value = "task-1130246" [ 988.973488] env[61991]: _type = "Task" [ 988.973488] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.979826] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.980032] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.981016] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3da97698-2ed0-46c5-9bc9-a1ab3d5cc539 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.986717] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 988.986717] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5227e6db-5bf9-7d93-da44-64bc6818ccc9" [ 988.986717] env[61991]: _type = "Task" [ 988.986717] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.989943] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130246, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.000869] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5227e6db-5bf9-7d93-da44-64bc6818ccc9, 'name': SearchDatastore_Task, 'duration_secs': 0.008172} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.001559] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d327d779-42f0-45ef-837b-34277fc70861 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.006257] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7604505-95c0-4b01-9e3e-5891f80e3cfe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.009435] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 989.009435] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5253a5cd-d7c7-6912-9ea5-075d1fdeed28" [ 989.009435] env[61991]: _type = "Task" [ 989.009435] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.015074] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823b3909-9fc2-4f36-94c5-e0d1a1670fc8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.022625] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5253a5cd-d7c7-6912-9ea5-075d1fdeed28, 'name': SearchDatastore_Task, 'duration_secs': 0.007741} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.025128] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.025532] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 679d3629-63e0-452f-9cee-1da4b9b843de/679d3629-63e0-452f-9cee-1da4b9b843de.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.025881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.026158] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.026423] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd192711-58c9-43f3-8498-b2b17876970a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.055581] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ece5eb3d-63e0-4e07-b51e-4e9603f031f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.058933] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66cd5f6-87dc-4a34-832a-d98a9a3549e1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.062940] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 989.062940] env[61991]: value = "task-1130247" [ 989.062940] env[61991]: _type = "Task" [ 989.062940] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.070630] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf547798-a575-4666-8cc3-717496dcec33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.074301] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.074478] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.078047] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71031825-fa1b-4f6a-849a-bbcb36854da8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.080295] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.089516] env[61991]: DEBUG nova.compute.provider_tree [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.092337] env[61991]: DEBUG nova.network.neutron [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.093492] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 989.093492] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db195c-7f15-c2c7-2f22-137188391895" [ 989.093492] env[61991]: _type = "Task" [ 989.093492] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.104443] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db195c-7f15-c2c7-2f22-137188391895, 'name': SearchDatastore_Task, 'duration_secs': 0.00765} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.104443] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dcc2a34-51dd-4598-875a-addea9b9bdfe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.112894] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 989.112894] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ecd275-c8b8-f7d9-eac3-0b24a2b7cf78" [ 989.112894] env[61991]: _type = "Task" [ 989.112894] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.121595] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ecd275-c8b8-f7d9-eac3-0b24a2b7cf78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.329884] env[61991]: DEBUG oslo_concurrency.lockutils [req-34a1434f-4c24-4136-9e42-9b93b38b361c req-c3667810-bbb5-4ba5-a2a1-1358d081414a service nova] Releasing lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.330482] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 989.404103] env[61991]: DEBUG oslo_vmware.api [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Task: {'id': task-1130245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14076} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.404437] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.404657] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.404841] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.405068] env[61991]: INFO nova.compute.manager [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 989.405350] env[61991]: DEBUG oslo.service.loopingcall [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.405580] env[61991]: DEBUG nova.compute.manager [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.405692] env[61991]: DEBUG nova.network.neutron [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 989.490081] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130246, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440069} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.490081] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 100d5fc0-e9d9-4892-9f60-99d657c533ed/100d5fc0-e9d9-4892-9f60-99d657c533ed.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.490476] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.490476] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6c1f986-84d4-4593-aa04-b226f80a5a11 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.497309] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 989.497309] env[61991]: value = "task-1130248" [ 989.497309] env[61991]: _type = "Task" [ 989.497309] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.509052] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130248, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.577516] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130247, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.598414] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.624152] env[61991]: ERROR nova.scheduler.client.report [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [req-ac3a4fbf-2130-4f15-bfab-c5c8545f6eff] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ac3a4fbf-2130-4f15-bfab-c5c8545f6eff"}]} [ 989.630497] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ecd275-c8b8-f7d9-eac3-0b24a2b7cf78, 'name': SearchDatastore_Task, 'duration_secs': 0.06882} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.631281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.631735] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d/8005b8f6-6f4d-4f72-be5d-955fa114408d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.631823] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39fbef01-7e50-4079-aa71-b183ddd7ee6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.641926] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 989.641926] env[61991]: value = "task-1130249" [ 989.641926] env[61991]: _type = "Task" [ 989.641926] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.647327] env[61991]: DEBUG nova.scheduler.client.report [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 989.656366] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.672315] env[61991]: DEBUG nova.scheduler.client.report [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 989.672595] env[61991]: DEBUG nova.compute.provider_tree [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.685626] env[61991]: DEBUG nova.scheduler.client.report [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 989.711264] env[61991]: DEBUG nova.scheduler.client.report [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 989.775303] env[61991]: DEBUG nova.compute.manager [req-ca97eccf-49f9-48d5-a43d-85e6b0b7d403 req-e607017a-295b-4d89-b0d9-5c19008ce4dc service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Received event network-vif-deleted-c7263e46-a164-4813-be3e-faae7fc31ff2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.775570] env[61991]: INFO nova.compute.manager [req-ca97eccf-49f9-48d5-a43d-85e6b0b7d403 req-e607017a-295b-4d89-b0d9-5c19008ce4dc service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Neutron deleted interface c7263e46-a164-4813-be3e-faae7fc31ff2; detaching it from the instance and deleting it from the info cache [ 989.775671] env[61991]: DEBUG nova.network.neutron [req-ca97eccf-49f9-48d5-a43d-85e6b0b7d403 req-e607017a-295b-4d89-b0d9-5c19008ce4dc service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.848217] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.916859] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e1fc2f-d2af-49c6-b45b-99b7cedaf157 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.924625] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d32d37f-0e29-434a-adc8-f72d34517ddf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.955364] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c298454-6fa3-43df-9a35-8a31a4803427 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.962274] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0171ac-feb6-424c-b228-597a64cebca7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.975177] env[61991]: DEBUG nova.compute.provider_tree [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 990.005577] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130248, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08803} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.005791] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.006651] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fa88eb-0e2f-48ec-80c1-d93e772a5f82 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.027150] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 100d5fc0-e9d9-4892-9f60-99d657c533ed/100d5fc0-e9d9-4892-9f60-99d657c533ed.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.027403] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9386356e-5425-4b47-8ad8-acbc22079eb2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.046306] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 990.046306] env[61991]: value = "task-1130250" [ 990.046306] env[61991]: _type = "Task" [ 990.046306] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.053664] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130250, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.072533] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130247, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.151702] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.254565] env[61991]: DEBUG nova.network.neutron [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.278344] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-971c6e4f-e26a-43f3-b935-5c361dd03a56 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.287370] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ecc583-55b3-4726-821d-0c3c49c6bf2a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.318689] env[61991]: DEBUG nova.compute.manager [req-ca97eccf-49f9-48d5-a43d-85e6b0b7d403 req-e607017a-295b-4d89-b0d9-5c19008ce4dc service nova] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Detach interface failed, port_id=c7263e46-a164-4813-be3e-faae7fc31ff2, reason: Instance 56188232-bed0-4a4b-a4bc-01edbb85cbe4 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 990.508530] env[61991]: DEBUG nova.scheduler.client.report [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 145 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 990.508832] env[61991]: DEBUG nova.compute.provider_tree [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 145 to 146 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 990.509063] env[61991]: DEBUG nova.compute.provider_tree [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 990.556619] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130250, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.573477] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130247, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.651332] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.757228] env[61991]: INFO nova.compute.manager [-] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Took 1.35 seconds to deallocate network for instance. [ 991.014680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.262s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.015197] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.018949] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.604s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.060929] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130250, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.075343] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130247, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.68781} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.075615] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 679d3629-63e0-452f-9cee-1da4b9b843de/679d3629-63e0-452f-9cee-1da4b9b843de.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.075839] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.076112] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-166c1330-155a-486c-92c6-f30740bbb442 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.084463] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 991.084463] env[61991]: value = "task-1130251" [ 991.084463] env[61991]: _type = "Task" [ 991.084463] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.092706] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130251, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.114435] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f3e0a7-064c-4ef8-80b5-d105b8037423 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.132042] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.150235] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130249, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.470055} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.150480] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d/8005b8f6-6f4d-4f72-be5d-955fa114408d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.150693] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.150929] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be70bcd0-1703-4fdf-ac18-fd867d756f37 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.156369] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 991.156369] env[61991]: value = "task-1130252" [ 991.156369] env[61991]: _type = "Task" [ 991.156369] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.163605] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130252, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.264253] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.522225] env[61991]: DEBUG nova.compute.utils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.525695] env[61991]: INFO nova.compute.claims [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.528970] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.529150] env[61991]: DEBUG nova.network.neutron [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 991.557855] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130250, 'name': ReconfigVM_Task, 'duration_secs': 1.243545} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.558138] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 100d5fc0-e9d9-4892-9f60-99d657c533ed/100d5fc0-e9d9-4892-9f60-99d657c533ed.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.558745] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0bb6cceb-7a57-4243-86e1-db38333e3c45 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.564974] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 991.564974] env[61991]: value = "task-1130253" [ 991.564974] env[61991]: _type = "Task" [ 991.564974] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.572772] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130253, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.580357] env[61991]: DEBUG nova.policy [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8c995dbd884e95bd04e034e2999a56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e4cb3d1518d498a8cdc2aee3acb90cb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.593263] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130251, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.259322} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.593512] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.594274] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992b85c1-ea74-4464-8efb-d16f96247ea4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.614022] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 679d3629-63e0-452f-9cee-1da4b9b843de/679d3629-63e0-452f-9cee-1da4b9b843de.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.614292] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd8e594e-e9ba-4733-9b4e-2386e6729409 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.634846] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 991.634846] env[61991]: value = "task-1130254" [ 991.634846] env[61991]: _type = "Task" [ 991.634846] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.639551] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.639823] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5dda469-c5ab-46db-8bbf-554cd1164724 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.645939] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130254, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.647212] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 991.647212] env[61991]: value = "task-1130255" [ 991.647212] env[61991]: _type = "Task" [ 991.647212] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.654612] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.664129] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130252, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078419} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.664437] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.665226] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d458ef-fda4-4bc7-a919-b16d0b9d8e5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.687343] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d/8005b8f6-6f4d-4f72-be5d-955fa114408d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.687658] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25ef1b79-6b9c-4767-a900-a2e248882207 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.706572] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 991.706572] env[61991]: value = "task-1130256" [ 991.706572] env[61991]: _type = "Task" [ 991.706572] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.715930] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130256, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.883660] env[61991]: DEBUG nova.network.neutron [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Successfully created port: f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.032035] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.036410] env[61991]: INFO nova.compute.resource_tracker [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating resource usage from migration 09b55899-bef6-4618-9ecd-0c7519b9cac5 [ 992.077200] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130253, 'name': Rename_Task, 'duration_secs': 0.185041} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.077626] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.077795] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9afe750-d356-444f-adac-a240b451aedb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.083552] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 992.083552] env[61991]: value = "task-1130257" [ 992.083552] env[61991]: _type = "Task" [ 992.083552] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.093404] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130257, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.148622] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130254, 'name': ReconfigVM_Task, 'duration_secs': 0.337797} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.151145] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 679d3629-63e0-452f-9cee-1da4b9b843de/679d3629-63e0-452f-9cee-1da4b9b843de.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.153723] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4234986-b0ec-48c9-b98c-413dcdc89b34 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.159534] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.160734] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 992.160734] env[61991]: value = "task-1130258" [ 992.160734] env[61991]: _type = "Task" [ 992.160734] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.170149] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130258, 'name': Rename_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.219036] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130256, 'name': ReconfigVM_Task, 'duration_secs': 0.335803} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.221499] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d/8005b8f6-6f4d-4f72-be5d-955fa114408d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.222444] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a8a5c34-8a9d-47d8-8bc7-f8801ccfbcac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.228656] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 992.228656] env[61991]: value = "task-1130259" [ 992.228656] env[61991]: _type = "Task" [ 992.228656] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.239391] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130259, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.245398] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c34dc2-40f2-4b5e-8222-6eb2893ca755 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.253184] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc8c3c3-6077-4416-9adc-3537efbe1b23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.287613] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7495da1c-dbde-4441-9aaf-655bebc7cb8d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.295569] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e78394a-8180-4533-9a7d-670cc9966ba1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.310664] env[61991]: DEBUG nova.compute.provider_tree [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 992.593086] env[61991]: DEBUG oslo_vmware.api [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130257, 'name': PowerOnVM_Task, 'duration_secs': 0.480926} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.593363] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.593568] env[61991]: INFO nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Took 9.81 seconds to spawn the instance on the hypervisor. [ 992.593751] env[61991]: DEBUG nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.594509] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76304ff8-61d0-4376-a0ad-0f3c1f323408 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.656790] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130255, 'name': PowerOffVM_Task, 'duration_secs': 0.834057} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.657035] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.657229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.669695] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130258, 'name': Rename_Task, 'duration_secs': 0.14552} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.669955] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.670198] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46cf83d1-9cb8-46bd-87d7-ace46ea96697 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.675938] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 992.675938] env[61991]: value = "task-1130260" [ 992.675938] env[61991]: _type = "Task" [ 992.675938] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.683829] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130260, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.738378] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130259, 'name': Rename_Task, 'duration_secs': 0.146979} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.738670] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.738909] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2e03e06-f9fa-4d43-bb88-7dec25582d03 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.745066] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 992.745066] env[61991]: value = "task-1130261" [ 992.745066] env[61991]: _type = "Task" [ 992.745066] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.753582] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130261, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.842949] env[61991]: DEBUG nova.scheduler.client.report [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 992.843319] env[61991]: DEBUG nova.compute.provider_tree [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 146 to 147 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 992.843514] env[61991]: DEBUG nova.compute.provider_tree [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 993.046686] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.074338] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.074656] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.074823] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.075016] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.075171] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.075320] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.075533] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.075735] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.075856] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.076031] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.076211] env[61991]: DEBUG nova.virt.hardware [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.077172] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e585f3-84ef-45d8-84ec-70664c0c3460 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.085079] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e7f3c3-2619-4113-b85a-c6974c885582 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.115403] env[61991]: INFO nova.compute.manager [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Took 21.22 seconds to build instance. [ 993.163267] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.163553] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.163727] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.163951] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.164076] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.164259] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.164526] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.164655] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.164820] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.164985] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.165181] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.170570] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3eb4a13e-7d74-428d-9ed5-5afc9498fea8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.189900] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130260, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.191524] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 993.191524] env[61991]: value = "task-1130262" [ 993.191524] env[61991]: _type = "Task" [ 993.191524] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.199839] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.255256] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130261, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.348975] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.331s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.349219] env[61991]: INFO nova.compute.manager [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Migrating [ 993.355923] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.508s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.357793] env[61991]: INFO nova.compute.claims [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.493532] env[61991]: DEBUG nova.compute.manager [req-c0f12251-0013-407a-8595-cb4348d93de5 req-bf6d1a4f-e536-45e8-ace3-8d04f4ebd1b5 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Received event network-vif-plugged-f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.493810] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0f12251-0013-407a-8595-cb4348d93de5 req-bf6d1a4f-e536-45e8-ace3-8d04f4ebd1b5 service nova] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.494185] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0f12251-0013-407a-8595-cb4348d93de5 req-bf6d1a4f-e536-45e8-ace3-8d04f4ebd1b5 service nova] Lock "5be79140-21f9-493c-a840-2d905b8d66bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.494376] env[61991]: DEBUG oslo_concurrency.lockutils [req-c0f12251-0013-407a-8595-cb4348d93de5 req-bf6d1a4f-e536-45e8-ace3-8d04f4ebd1b5 service nova] Lock "5be79140-21f9-493c-a840-2d905b8d66bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.494599] env[61991]: DEBUG nova.compute.manager [req-c0f12251-0013-407a-8595-cb4348d93de5 req-bf6d1a4f-e536-45e8-ace3-8d04f4ebd1b5 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] No waiting events found dispatching network-vif-plugged-f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 993.494820] env[61991]: WARNING nova.compute.manager [req-c0f12251-0013-407a-8595-cb4348d93de5 req-bf6d1a4f-e536-45e8-ace3-8d04f4ebd1b5 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Received unexpected event network-vif-plugged-f5641235-5202-4d60-a351-b93be008988e for instance with vm_state building and task_state spawning. [ 993.617625] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d88cbe75-0710-478e-b223-88a95448d467 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.728s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.658283] env[61991]: DEBUG nova.network.neutron [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Successfully updated port: f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.690376] env[61991]: DEBUG oslo_vmware.api [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130260, 'name': PowerOnVM_Task, 'duration_secs': 0.854869} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.690632] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.690837] env[61991]: INFO nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Took 6.26 seconds to spawn the instance on the hypervisor. [ 993.691059] env[61991]: DEBUG nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.691877] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726ab087-a033-4825-a907-233a3aff1ea9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.705975] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130262, 'name': ReconfigVM_Task, 'duration_secs': 0.17727} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.706314] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.757036] env[61991]: DEBUG oslo_vmware.api [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130261, 'name': PowerOnVM_Task, 'duration_secs': 0.771586} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.757241] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.757445] env[61991]: INFO nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Took 8.66 seconds to spawn the instance on the hypervisor. [ 993.757629] env[61991]: DEBUG nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.758433] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a99043b-370d-4af9-a65b-5377e00afd89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.871939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.872237] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.872419] env[61991]: DEBUG nova.network.neutron [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.160746] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.160954] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.161217] env[61991]: DEBUG nova.network.neutron [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.215853] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.216223] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.216223] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.218271] env[61991]: DEBUG nova.virt.hardware [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.222763] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Reconfiguring VM instance instance-00000046 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.224873] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1a697a8-bf3b-46c9-a7f6-26df0a7758ef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.238856] env[61991]: INFO nova.compute.manager [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Took 18.05 seconds to build instance. [ 994.246795] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 994.246795] env[61991]: value = "task-1130263" [ 994.246795] env[61991]: _type = "Task" [ 994.246795] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.255196] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130263, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.275267] env[61991]: INFO nova.compute.manager [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Took 20.97 seconds to build instance. [ 994.576530] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d05c47-aa37-498c-9e8a-841b1f1da9d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.586563] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5db4477-cba5-448b-b249-2ad2e5571e02 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.623255] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2621bed1-c379-4430-beed-8d786c7dbf90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.635047] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e809b04-609a-423e-becb-71c68f2197e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.650597] env[61991]: DEBUG nova.compute.provider_tree [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.680462] env[61991]: DEBUG nova.network.neutron [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.713111] env[61991]: DEBUG nova.network.neutron [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.741557] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b6efe06f-7c5d-4f4b-bb84-1e08a2dba34e tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "679d3629-63e0-452f-9cee-1da4b9b843de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.562s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.757136] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.777336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4064b35b-175e-4c9e-833b-0ae6b8c51939 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.485s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.903434] env[61991]: DEBUG nova.network.neutron [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updating instance_info_cache with network_info: [{"id": "f5641235-5202-4d60-a351-b93be008988e", "address": "fa:16:3e:e2:ec:86", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5641235-52", "ovs_interfaceid": "f5641235-5202-4d60-a351-b93be008988e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.156023] env[61991]: DEBUG nova.scheduler.client.report [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.183346] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.257516] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.406248] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.406644] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Instance network_info: |[{"id": "f5641235-5202-4d60-a351-b93be008988e", "address": "fa:16:3e:e2:ec:86", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5641235-52", "ovs_interfaceid": "f5641235-5202-4d60-a351-b93be008988e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.407143] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:ec:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5641235-5202-4d60-a351-b93be008988e', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.416759] env[61991]: DEBUG oslo.service.loopingcall [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.417018] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 995.417291] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ef675a3-51d8-47ae-8417-5dbaf95b1fe3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.438134] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.438134] env[61991]: value = "task-1130265" [ 995.438134] env[61991]: _type = "Task" [ 995.438134] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.446505] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130265, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.514410] env[61991]: DEBUG nova.compute.manager [None req-4ca41661-28f1-4e55-bb80-db3a6cce4048 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.515819] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07453866-bf1d-4734-a928-a838128529ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.569515] env[61991]: DEBUG nova.compute.manager [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 995.577824] env[61991]: DEBUG nova.compute.manager [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Received event network-changed-96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.578056] env[61991]: DEBUG nova.compute.manager [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Refreshing instance network info cache due to event network-changed-96d11c6e-409f-4baf-983b-c066592aa09b. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.578313] env[61991]: DEBUG oslo_concurrency.lockutils [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] Acquiring lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.578480] env[61991]: DEBUG oslo_concurrency.lockutils [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] Acquired lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.578651] env[61991]: DEBUG nova.network.neutron [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Refreshing network info cache for port 96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 995.585076] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "679d3629-63e0-452f-9cee-1da4b9b843de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.585349] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "679d3629-63e0-452f-9cee-1da4b9b843de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.585727] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "679d3629-63e0-452f-9cee-1da4b9b843de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.586012] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "679d3629-63e0-452f-9cee-1da4b9b843de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.586219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "679d3629-63e0-452f-9cee-1da4b9b843de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.588773] env[61991]: INFO nova.compute.manager [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Terminating instance [ 995.590473] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "refresh_cache-679d3629-63e0-452f-9cee-1da4b9b843de" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.590627] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquired lock "refresh_cache-679d3629-63e0-452f-9cee-1da4b9b843de" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.590870] env[61991]: DEBUG nova.network.neutron [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 995.660281] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.660819] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 995.663320] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.399s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.664026] env[61991]: DEBUG nova.objects.instance [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lazy-loading 'resources' on Instance uuid 56188232-bed0-4a4b-a4bc-01edbb85cbe4 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.757920] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130263, 'name': ReconfigVM_Task, 'duration_secs': 1.208968} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.758235] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Reconfigured VM instance instance-00000046 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.758988] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab22bddc-3c8e-47aa-9f95-9dc3b40e1bce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.781854] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e/d36bd981-9ea3-46f6-8376-ac1e0c3bf61e.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.782130] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f840250c-1d7a-4703-8725-1c7eb92dcb93 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.799171] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 995.799171] env[61991]: value = "task-1130266" [ 995.799171] env[61991]: _type = "Task" [ 995.799171] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.806535] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130266, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.950703] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130265, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.027305] env[61991]: INFO nova.compute.manager [None req-4ca41661-28f1-4e55-bb80-db3a6cce4048 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] instance snapshotting [ 996.027527] env[61991]: DEBUG nova.objects.instance [None req-4ca41661-28f1-4e55-bb80-db3a6cce4048 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lazy-loading 'flavor' on Instance uuid 679d3629-63e0-452f-9cee-1da4b9b843de {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.094906] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.112928] env[61991]: DEBUG nova.network.neutron [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.166869] env[61991]: DEBUG nova.compute.utils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.175688] env[61991]: DEBUG nova.network.neutron [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.176482] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 996.176841] env[61991]: DEBUG nova.network.neutron [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 996.242946] env[61991]: DEBUG nova.policy [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '169b21c34f2f48cb83d848c13efb6ac7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '860b2205deb34624b529effcc1f0b44f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.310921] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130266, 'name': ReconfigVM_Task, 'duration_secs': 0.258619} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.311268] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Reconfigured VM instance instance-00000046 to attach disk [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e/d36bd981-9ea3-46f6-8376-ac1e0c3bf61e.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.311557] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 996.334336] env[61991]: DEBUG nova.network.neutron [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updated VIF entry in instance network info cache for port 96d11c6e-409f-4baf-983b-c066592aa09b. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 996.334751] env[61991]: DEBUG nova.network.neutron [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updating instance_info_cache with network_info: [{"id": "96d11c6e-409f-4baf-983b-c066592aa09b", "address": "fa:16:3e:76:45:34", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96d11c6e-40", "ovs_interfaceid": "96d11c6e-409f-4baf-983b-c066592aa09b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.409244] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce25996-4243-4277-8c1f-a84e2369c041 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.416509] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb6e6dd-6a31-4366-8cc5-8245ed542600 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.450516] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34b918a-b1e9-44ab-8561-f47fe1448b84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.457677] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130265, 'name': CreateVM_Task, 'duration_secs': 0.990105} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.459625] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.460401] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.460579] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.460909] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.462114] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05864969-3a74-4215-bda7-f7b956fd9e47 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.465964] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fe3141d-2fac-4693-b224-b4d22c71f619 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.480033] env[61991]: DEBUG nova.compute.provider_tree [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.486058] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 996.486058] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52316da1-68bc-f266-73fb-29de0d83e176" [ 996.486058] env[61991]: _type = "Task" [ 996.486058] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.495586] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52316da1-68bc-f266-73fb-29de0d83e176, 'name': SearchDatastore_Task, 'duration_secs': 0.009964} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.495896] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.497267] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.497267] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.497267] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.497267] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.497267] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e5d52c1-ece1-4caa-9ae0-853afbd9b80a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.504762] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.504953] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.505683] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe083891-9331-4d09-802b-ec3d674797aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.508528] env[61991]: DEBUG nova.network.neutron [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Successfully created port: 16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.513533] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 996.513533] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5226514d-b39b-d455-7876-b66c32e04c22" [ 996.513533] env[61991]: _type = "Task" [ 996.513533] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.521225] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5226514d-b39b-d455-7876-b66c32e04c22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.533781] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b2b3f7-b89c-427e-922a-c7115f2a3bf8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.551512] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fa1860-5e42-4b43-a3d9-b8ba754e71f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.673585] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 996.688061] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Releasing lock "refresh_cache-679d3629-63e0-452f-9cee-1da4b9b843de" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.688061] env[61991]: DEBUG nova.compute.manager [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.688061] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.688061] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4d3338-16b5-4f55-bc34-a15a059dfd26 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.698394] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.699448] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c094e83e-78e2-43e2-85ac-fd9f321aa355 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.701842] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61013205-1910-4be5-96f5-b05309c1a379 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.728437] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 996.733846] env[61991]: DEBUG oslo_vmware.api [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 996.733846] env[61991]: value = "task-1130267" [ 996.733846] env[61991]: _type = "Task" [ 996.733846] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.744247] env[61991]: DEBUG oslo_vmware.api [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130267, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.819055] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2f42f5-af6c-4677-8285-e1bb55cc1a8f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.839674] env[61991]: DEBUG oslo_concurrency.lockutils [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] Releasing lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.840058] env[61991]: DEBUG nova.compute.manager [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Received event network-changed-f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.840314] env[61991]: DEBUG nova.compute.manager [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Refreshing instance network info cache due to event network-changed-f5641235-5202-4d60-a351-b93be008988e. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.840579] env[61991]: DEBUG oslo_concurrency.lockutils [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] Acquiring lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.840756] env[61991]: DEBUG oslo_concurrency.lockutils [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] Acquired lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.840960] env[61991]: DEBUG nova.network.neutron [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Refreshing network info cache for port f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 996.842780] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd0589d-6ccb-471c-b44c-66fce308e170 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.862668] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 996.984516] env[61991]: DEBUG nova.scheduler.client.report [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.024842] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5226514d-b39b-d455-7876-b66c32e04c22, 'name': SearchDatastore_Task, 'duration_secs': 0.008052} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.027414] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9810aa3f-9698-4f09-809e-251556452d64 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.030867] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 997.030867] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5207667b-592e-da61-d1cb-2eaad264a2ba" [ 997.030867] env[61991]: _type = "Task" [ 997.030867] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.039417] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5207667b-592e-da61-d1cb-2eaad264a2ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.060226] env[61991]: DEBUG nova.compute.manager [None req-4ca41661-28f1-4e55-bb80-db3a6cce4048 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Instance disappeared during snapshot {{(pid=61991) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 997.201264] env[61991]: DEBUG nova.compute.manager [None req-4ca41661-28f1-4e55-bb80-db3a6cce4048 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Found 0 images (rotation: 2) {{(pid=61991) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 997.236450] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.236791] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25491102-e521-4dda-8475-371acb879b4b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.248604] env[61991]: DEBUG oslo_vmware.api [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130267, 'name': PowerOffVM_Task, 'duration_secs': 0.313811} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.250225] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.250518] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.250966] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 997.250966] env[61991]: value = "task-1130268" [ 997.250966] env[61991]: _type = "Task" [ 997.250966] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.251315] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe6aaffa-6b2e-42b1-ae2a-5469bdff7f95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.261384] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130268, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.277107] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.277359] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.277540] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Deleting the datastore file [datastore1] 679d3629-63e0-452f-9cee-1da4b9b843de {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.277834] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcec8b03-81e5-4646-8981-af3dfbbe6c1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.283540] env[61991]: DEBUG oslo_vmware.api [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for the task: (returnval){ [ 997.283540] env[61991]: value = "task-1130270" [ 997.283540] env[61991]: _type = "Task" [ 997.283540] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.292821] env[61991]: DEBUG oslo_vmware.api [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130270, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.426136] env[61991]: DEBUG nova.network.neutron [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Port 9452191d-e715-4ad9-a48e-43aebb90b551 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 997.487851] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.824s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.490626] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.396s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.526738] env[61991]: INFO nova.scheduler.client.report [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Deleted allocations for instance 56188232-bed0-4a4b-a4bc-01edbb85cbe4 [ 997.549423] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5207667b-592e-da61-d1cb-2eaad264a2ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009204} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.549793] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.550114] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 5be79140-21f9-493c-a840-2d905b8d66bf/5be79140-21f9-493c-a840-2d905b8d66bf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.553246] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa1c39f1-c24b-4434-b905-59bb976a4aca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.569811] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 997.569811] env[61991]: value = "task-1130271" [ 997.569811] env[61991]: _type = "Task" [ 997.569811] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.586785] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130271, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.597951] env[61991]: DEBUG nova.network.neutron [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updated VIF entry in instance network info cache for port f5641235-5202-4d60-a351-b93be008988e. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 997.598439] env[61991]: DEBUG nova.network.neutron [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updating instance_info_cache with network_info: [{"id": "f5641235-5202-4d60-a351-b93be008988e", "address": "fa:16:3e:e2:ec:86", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5641235-52", "ovs_interfaceid": "f5641235-5202-4d60-a351-b93be008988e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.683178] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 997.706686] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.706946] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.707121] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.707326] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.707531] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.707697] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.707916] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.708090] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.708263] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.708431] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.708608] env[61991]: DEBUG nova.virt.hardware [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.709511] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc1203f-7dc2-413a-a628-76e0d3bddca3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.718403] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b36133e-ce0d-4cdf-8658-bb04dad0408e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.761492] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130268, 'name': PowerOffVM_Task, 'duration_secs': 0.379822} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.761747] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.761930] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 997.793288] env[61991]: DEBUG oslo_vmware.api [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Task: {'id': task-1130270, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099735} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.793551] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.793742] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.793917] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.794113] env[61991]: INFO nova.compute.manager [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Took 1.11 seconds to destroy the instance on the hypervisor. [ 997.794442] env[61991]: DEBUG oslo.service.loopingcall [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.794663] env[61991]: DEBUG nova.compute.manager [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.794777] env[61991]: DEBUG nova.network.neutron [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 997.810313] env[61991]: DEBUG nova.network.neutron [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 997.998113] env[61991]: INFO nova.compute.claims [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.035551] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f1a9db77-c75b-41db-a336-cb00b36a71d8 tempest-ServersNegativeTestJSON-200394740 tempest-ServersNegativeTestJSON-200394740-project-member] Lock "56188232-bed0-4a4b-a4bc-01edbb85cbe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.750s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.080428] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130271, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.102371] env[61991]: DEBUG oslo_concurrency.lockutils [req-f49134cc-e3a2-4b6b-af5d-03b8fcffb137 req-12dc7e87-e22e-4b8b-a768-79e87af09850 service nova] Releasing lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.216080] env[61991]: DEBUG nova.compute.manager [req-ffbc9709-254e-4f81-81c4-b10c6cb00b98 req-e67b1a1f-476d-4d86-a546-81bb87c76fda service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Received event network-vif-plugged-16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.216508] env[61991]: DEBUG oslo_concurrency.lockutils [req-ffbc9709-254e-4f81-81c4-b10c6cb00b98 req-e67b1a1f-476d-4d86-a546-81bb87c76fda service nova] Acquiring lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.216676] env[61991]: DEBUG oslo_concurrency.lockutils [req-ffbc9709-254e-4f81-81c4-b10c6cb00b98 req-e67b1a1f-476d-4d86-a546-81bb87c76fda service nova] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.216960] env[61991]: DEBUG oslo_concurrency.lockutils [req-ffbc9709-254e-4f81-81c4-b10c6cb00b98 req-e67b1a1f-476d-4d86-a546-81bb87c76fda service nova] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.217168] env[61991]: DEBUG nova.compute.manager [req-ffbc9709-254e-4f81-81c4-b10c6cb00b98 req-e67b1a1f-476d-4d86-a546-81bb87c76fda service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] No waiting events found dispatching network-vif-plugged-16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 998.217344] env[61991]: WARNING nova.compute.manager [req-ffbc9709-254e-4f81-81c4-b10c6cb00b98 req-e67b1a1f-476d-4d86-a546-81bb87c76fda service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Received unexpected event network-vif-plugged-16df29ee-00a7-4144-8e1b-3b0c0699817e for instance with vm_state building and task_state spawning. [ 998.268189] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.268433] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.268625] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.268860] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.268980] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.269148] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.269360] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.269564] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.269744] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.269913] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.270102] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.275127] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-329243f3-d644-4b45-9a3c-73c5eaecc109 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.291504] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 998.291504] env[61991]: value = "task-1130273" [ 998.291504] env[61991]: _type = "Task" [ 998.291504] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.304802] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130273, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.314282] env[61991]: DEBUG nova.network.neutron [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.398827] env[61991]: DEBUG nova.network.neutron [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Successfully updated port: 16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.454422] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.454673] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.454851] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.508274] env[61991]: INFO nova.compute.resource_tracker [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating resource usage from migration 5a1d3705-160c-468f-a272-eff9cc47f565 [ 998.582400] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130271, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625866} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.585241] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 5be79140-21f9-493c-a840-2d905b8d66bf/5be79140-21f9-493c-a840-2d905b8d66bf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.585489] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.585977] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e765b1b0-5385-40bb-a58f-757cbe1b8c06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.593317] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 998.593317] env[61991]: value = "task-1130274" [ 998.593317] env[61991]: _type = "Task" [ 998.593317] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.605364] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130274, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.728151] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e571411b-dff1-4ec6-be44-2682ba568db0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.746856] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bc220a-f41d-48c2-a2de-16ef8f5f97dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.803615] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a05ed8d-6d9d-4f06-80f8-7c08961c41e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.815990] env[61991]: INFO nova.compute.manager [-] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Took 1.02 seconds to deallocate network for instance. [ 998.817252] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.824067] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6f737c-a30b-405b-9ed7-a418d895f7da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.852033] env[61991]: DEBUG nova.compute.provider_tree [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.902217] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.902562] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquired lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.902752] env[61991]: DEBUG nova.network.neutron [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.104657] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130274, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.318745] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.338050] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.355621] env[61991]: DEBUG nova.scheduler.client.report [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.464979] env[61991]: DEBUG nova.network.neutron [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 999.508564] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.508843] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.509090] env[61991]: DEBUG nova.network.neutron [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.611029] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130274, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.786265} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.611029] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.611029] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12773d65-41a0-4761-a59b-2643b3950724 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.645854] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 5be79140-21f9-493c-a840-2d905b8d66bf/5be79140-21f9-493c-a840-2d905b8d66bf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.646677] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67f14b0a-555a-47c2-8bde-97c2ff8664a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.663103] env[61991]: DEBUG nova.network.neutron [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Updating instance_info_cache with network_info: [{"id": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "address": "fa:16:3e:50:57:72", "network": {"id": "ca86d109-d0dd-490f-be31-686b0fed7f9e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2065252436-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "860b2205deb34624b529effcc1f0b44f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae215ba8-f7a5-4b23-a055-90316d29817f", "external-id": "nsx-vlan-transportzone-798", "segmentation_id": 798, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16df29ee-00", "ovs_interfaceid": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.670533] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 999.670533] env[61991]: value = "task-1130275" [ 999.670533] env[61991]: _type = "Task" [ 999.670533] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.683921] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130275, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.817871] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130273, 'name': ReconfigVM_Task, 'duration_secs': 1.19483} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.819117] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 999.862304] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.371s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.862858] env[61991]: INFO nova.compute.manager [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Migrating [ 999.874748] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.537s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.875679] env[61991]: DEBUG nova.objects.instance [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lazy-loading 'resources' on Instance uuid 679d3629-63e0-452f-9cee-1da4b9b843de {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.166406] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Releasing lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.166851] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Instance network_info: |[{"id": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "address": "fa:16:3e:50:57:72", "network": {"id": "ca86d109-d0dd-490f-be31-686b0fed7f9e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2065252436-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "860b2205deb34624b529effcc1f0b44f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae215ba8-f7a5-4b23-a055-90316d29817f", "external-id": "nsx-vlan-transportzone-798", "segmentation_id": 798, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16df29ee-00", "ovs_interfaceid": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1000.167410] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:57:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae215ba8-f7a5-4b23-a055-90316d29817f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16df29ee-00a7-4144-8e1b-3b0c0699817e', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.175344] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Creating folder: Project (860b2205deb34624b529effcc1f0b44f). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1000.176150] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f720cfb5-5fe0-4258-a845-871e6f0a4f0d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.189946] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130275, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.193034] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Created folder: Project (860b2205deb34624b529effcc1f0b44f) in parent group-v246753. [ 1000.193253] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Creating folder: Instances. Parent ref: group-v246959. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1000.193513] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81eff3ed-7082-441f-b94a-c59ba4652af0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.203420] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Created folder: Instances in parent group-v246959. [ 1000.203420] env[61991]: DEBUG oslo.service.loopingcall [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.203420] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.203420] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c87a5d39-cd77-49f2-b94a-c6b21a1cb8cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.223193] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.223193] env[61991]: value = "task-1130279" [ 1000.223193] env[61991]: _type = "Task" [ 1000.223193] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.233732] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130279, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.286621] env[61991]: DEBUG nova.compute.manager [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Received event network-changed-16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.286819] env[61991]: DEBUG nova.compute.manager [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Refreshing instance network info cache due to event network-changed-16df29ee-00a7-4144-8e1b-3b0c0699817e. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.287048] env[61991]: DEBUG oslo_concurrency.lockutils [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] Acquiring lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.287198] env[61991]: DEBUG oslo_concurrency.lockutils [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] Acquired lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.287429] env[61991]: DEBUG nova.network.neutron [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Refreshing network info cache for port 16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1000.315568] env[61991]: DEBUG nova.network.neutron [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.325858] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.326107] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.326268] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.326491] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.326650] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.326802] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.327012] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.327177] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.327343] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.327534] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.327672] env[61991]: DEBUG nova.virt.hardware [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.332862] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1000.333767] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0b8cde8-9249-4b65-af4e-4e72536a746c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.353273] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1000.353273] env[61991]: value = "task-1130280" [ 1000.353273] env[61991]: _type = "Task" [ 1000.353273] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.361880] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130280, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.394169] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.394568] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.394902] env[61991]: DEBUG nova.network.neutron [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.600562] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ba2b4d-d981-415b-a893-0866946a56ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.608168] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176b06fc-619a-4b81-b132-0a11c815dc06 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.640429] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17e8072-2dcd-4277-ae75-614a5f2f1630 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.647814] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b29b005-61bd-4d9b-bf53-4446ad8fc95f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.661199] env[61991]: DEBUG nova.compute.provider_tree [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.685314] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130275, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.733125] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130279, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.818249] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.864195] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130280, 'name': ReconfigVM_Task, 'duration_secs': 0.485908} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.867157] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1000.868059] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c0711f-0251-4b00-933f-6f7f52ce3bd5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.895280] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.898210] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceb88bcb-31db-4bd3-912d-e4df18710a17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.917613] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1000.917613] env[61991]: value = "task-1130281" [ 1000.917613] env[61991]: _type = "Task" [ 1000.917613] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.926191] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130281, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.068713] env[61991]: DEBUG nova.network.neutron [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Updated VIF entry in instance network info cache for port 16df29ee-00a7-4144-8e1b-3b0c0699817e. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1001.069191] env[61991]: DEBUG nova.network.neutron [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Updating instance_info_cache with network_info: [{"id": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "address": "fa:16:3e:50:57:72", "network": {"id": "ca86d109-d0dd-490f-be31-686b0fed7f9e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2065252436-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "860b2205deb34624b529effcc1f0b44f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae215ba8-f7a5-4b23-a055-90316d29817f", "external-id": "nsx-vlan-transportzone-798", "segmentation_id": 798, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16df29ee-00", "ovs_interfaceid": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.156374] env[61991]: DEBUG nova.network.neutron [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance_info_cache with network_info: [{"id": "15cc98a6-0818-4630-b737-e50e1fb71851", "address": "fa:16:3e:97:a6:51", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15cc98a6-08", "ovs_interfaceid": "15cc98a6-0818-4630-b737-e50e1fb71851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.164487] env[61991]: DEBUG nova.scheduler.client.report [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.186133] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130275, 'name': ReconfigVM_Task, 'duration_secs': 1.256964} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.186473] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 5be79140-21f9-493c-a840-2d905b8d66bf/5be79140-21f9-493c-a840-2d905b8d66bf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.187111] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1be21584-4aa9-49cf-83a1-77972412f9a8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.192878] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1001.192878] env[61991]: value = "task-1130282" [ 1001.192878] env[61991]: _type = "Task" [ 1001.192878] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.203553] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130282, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.233863] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130279, 'name': CreateVM_Task, 'duration_secs': 0.961759} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.234127] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.234767] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.234940] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.235319] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.235581] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-146eef84-a7cc-4002-9dc5-2d86ff45ee9f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.240508] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1001.240508] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a4206d-112b-1506-4336-266e39ee010d" [ 1001.240508] env[61991]: _type = "Task" [ 1001.240508] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.249181] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a4206d-112b-1506-4336-266e39ee010d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.344373] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbca973-804e-4b6f-a7d5-c59e5254c22c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.363183] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89741466-e402-4c29-b66a-6bc613faa818 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.370882] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.428225] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130281, 'name': ReconfigVM_Task, 'duration_secs': 0.411316} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.428510] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.428827] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.572518] env[61991]: DEBUG oslo_concurrency.lockutils [req-1e3a4a56-6c13-435b-ab13-11d9e4649162 req-25acb81a-9f4b-40aa-bb9b-1a3cfb4e64d7 service nova] Releasing lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.660072] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.669776] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.795s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.689044] env[61991]: INFO nova.scheduler.client.report [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Deleted allocations for instance 679d3629-63e0-452f-9cee-1da4b9b843de [ 1001.702658] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130282, 'name': Rename_Task, 'duration_secs': 0.143058} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.702918] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.703168] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d11052c3-ba8a-4700-9cb8-757a131f64d3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.709260] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1001.709260] env[61991]: value = "task-1130283" [ 1001.709260] env[61991]: _type = "Task" [ 1001.709260] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.716371] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.750552] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52a4206d-112b-1506-4336-266e39ee010d, 'name': SearchDatastore_Task, 'duration_secs': 0.010516} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.750868] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.751124] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.751365] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.751518] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.751696] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.751949] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-647ba95d-fcdd-41cf-ac8e-9f5c7d18f114 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.759630] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.759811] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.760517] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f47be66f-dfd3-466b-acfd-52aa15853d93 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.765404] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1001.765404] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526946b4-2584-82ba-d1ca-25d67f0f1abd" [ 1001.765404] env[61991]: _type = "Task" [ 1001.765404] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.773284] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526946b4-2584-82ba-d1ca-25d67f0f1abd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.878669] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.878996] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31cd6bca-4d2b-495b-8e3e-ffdafd7945e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.885577] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1001.885577] env[61991]: value = "task-1130285" [ 1001.885577] env[61991]: _type = "Task" [ 1001.885577] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.894612] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130285, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.935404] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0179d0cc-d21a-4ff4-a4b2-a9f81a41fbf9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.958744] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a419929e-fe22-4788-8952-d27f6cabc310 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.979016] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.197861] env[61991]: DEBUG oslo_concurrency.lockutils [None req-38169037-625f-42e4-ae0c-f509a750ac41 tempest-ServersAaction247Test-1744074558 tempest-ServersAaction247Test-1744074558-project-member] Lock "679d3629-63e0-452f-9cee-1da4b9b843de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.612s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.219111] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130283, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.276014] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526946b4-2584-82ba-d1ca-25d67f0f1abd, 'name': SearchDatastore_Task, 'duration_secs': 0.008166} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.277110] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-142ff806-1d06-4509-82bf-acf3e71d7469 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.282513] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1002.282513] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52408c05-e706-0d6c-a180-84ae7747077c" [ 1002.282513] env[61991]: _type = "Task" [ 1002.282513] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.290491] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52408c05-e706-0d6c-a180-84ae7747077c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.394753] env[61991]: DEBUG oslo_vmware.api [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130285, 'name': PowerOnVM_Task, 'duration_secs': 0.454644} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.395082] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.395238] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5d4481a4-367c-4dd1-967d-2b18308f3a82 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance 'd36bd981-9ea3-46f6-8376-ac1e0c3bf61e' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.719245] env[61991]: DEBUG oslo_vmware.api [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130283, 'name': PowerOnVM_Task, 'duration_secs': 0.551036} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.719568] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.720316] env[61991]: INFO nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Took 9.67 seconds to spawn the instance on the hypervisor. [ 1002.720521] env[61991]: DEBUG nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.721339] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5b9af9-8e24-4ee6-98d2-c84f8ee44022 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.793170] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52408c05-e706-0d6c-a180-84ae7747077c, 'name': SearchDatastore_Task, 'duration_secs': 0.009882} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.793487] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.793790] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6/39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.794090] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a9d749a-76bf-4ab8-b09d-dc5ce55d1927 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.800864] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1002.800864] env[61991]: value = "task-1130286" [ 1002.800864] env[61991]: _type = "Task" [ 1002.800864] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.808150] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130286, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.175696] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487064c5-b6a0-4e13-9db9-b339661a1d71 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.195322] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.238094] env[61991]: INFO nova.compute.manager [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Took 20.82 seconds to build instance. [ 1003.312279] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130286, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.642462] env[61991]: DEBUG nova.network.neutron [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Port 4e506178-06a6-49c6-a6b3-fa24ebafc988 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1003.701880] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.702207] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17917332-0b81-416f-ab88-3de2afd9207f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.709627] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1003.709627] env[61991]: value = "task-1130287" [ 1003.709627] env[61991]: _type = "Task" [ 1003.709627] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.717924] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.741011] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3fc6d620-9acc-4dda-ae98-0e5c451086f8 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.337s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.811317] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130286, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.744788} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.811576] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6/39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1003.811810] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.812081] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab32db9b-946c-4c42-877f-eadb63159541 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.819034] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1003.819034] env[61991]: value = "task-1130288" [ 1003.819034] env[61991]: _type = "Task" [ 1003.819034] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.827908] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130288, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.222748] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130287, 'name': PowerOffVM_Task, 'duration_secs': 0.182871} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.223155] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.223237] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1004.275789] env[61991]: DEBUG nova.compute.manager [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Received event network-changed-f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.275907] env[61991]: DEBUG nova.compute.manager [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Refreshing instance network info cache due to event network-changed-f5641235-5202-4d60-a351-b93be008988e. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.276168] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] Acquiring lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.276305] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] Acquired lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.276571] env[61991]: DEBUG nova.network.neutron [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Refreshing network info cache for port f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.330036] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130288, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07125} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.330286] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.331165] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bba97d4-d15e-4e3b-bc3f-986261506197 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.355415] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6/39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.356120] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9aaf38ae-6c0a-400b-a367-838ad03bed55 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.378139] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1004.378139] env[61991]: value = "task-1130290" [ 1004.378139] env[61991]: _type = "Task" [ 1004.378139] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.388793] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130290, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.463648] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.464037] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.464362] env[61991]: DEBUG nova.compute.manager [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Going to confirm migration 5 {{(pid=61991) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1004.666219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.666521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.666743] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.729973] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1004.730243] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1004.730411] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.730601] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1004.730751] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.730901] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1004.731130] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1004.731300] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1004.731486] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1004.731646] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1004.731825] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.737239] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0841064-eeda-44b0-bbe6-6b35f04bf18a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.752879] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1004.752879] env[61991]: value = "task-1130291" [ 1004.752879] env[61991]: _type = "Task" [ 1004.752879] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.761938] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130291, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.892344] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130290, 'name': ReconfigVM_Task, 'duration_secs': 0.432094} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.892799] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6/39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.893771] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-786bdc26-1bd2-4bc0-b2ff-b08668d92156 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.900658] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1004.900658] env[61991]: value = "task-1130292" [ 1004.900658] env[61991]: _type = "Task" [ 1004.900658] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.908951] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130292, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.007925] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.008287] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.008524] env[61991]: DEBUG nova.network.neutron [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.008769] env[61991]: DEBUG nova.objects.instance [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'info_cache' on Instance uuid d36bd981-9ea3-46f6-8376-ac1e0c3bf61e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.022898] env[61991]: DEBUG nova.network.neutron [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updated VIF entry in instance network info cache for port f5641235-5202-4d60-a351-b93be008988e. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1005.023432] env[61991]: DEBUG nova.network.neutron [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updating instance_info_cache with network_info: [{"id": "f5641235-5202-4d60-a351-b93be008988e", "address": "fa:16:3e:e2:ec:86", "network": {"id": "3c24826c-a656-4097-80b9-742681de1e68", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1401446213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9e4cb3d1518d498a8cdc2aee3acb90cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5641235-52", "ovs_interfaceid": "f5641235-5202-4d60-a351-b93be008988e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.262694] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130291, 'name': ReconfigVM_Task, 'duration_secs': 0.215637} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.263071] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.410371] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130292, 'name': Rename_Task, 'duration_secs': 0.15496} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.410606] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.410850] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33efd80d-e86f-4237-b97a-1cfcb83623f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.417106] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1005.417106] env[61991]: value = "task-1130293" [ 1005.417106] env[61991]: _type = "Task" [ 1005.417106] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.424138] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.527068] env[61991]: DEBUG oslo_concurrency.lockutils [req-2ea008ec-1922-4cc7-aad3-ba0b83fe13f1 req-09e1a06d-df3c-4322-984d-a8ddf4675932 service nova] Releasing lock "refresh_cache-5be79140-21f9-493c-a840-2d905b8d66bf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.696646] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.696867] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.697082] env[61991]: DEBUG nova.network.neutron [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.768972] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.769276] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.769385] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.769606] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.769780] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.769933] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.770159] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.770326] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.770492] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.770710] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.770900] env[61991]: DEBUG nova.virt.hardware [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.776120] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1005.776723] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d4836bc-149e-4d0c-91d4-259e2faca565 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.795498] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1005.795498] env[61991]: value = "task-1130294" [ 1005.795498] env[61991]: _type = "Task" [ 1005.795498] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.804819] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130294, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.928806] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130293, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.212674] env[61991]: DEBUG nova.network.neutron [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [{"id": "9452191d-e715-4ad9-a48e-43aebb90b551", "address": "fa:16:3e:9f:8a:89", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9452191d-e7", "ovs_interfaceid": "9452191d-e715-4ad9-a48e-43aebb90b551", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.305635] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130294, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.428983] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130293, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.455854] env[61991]: DEBUG nova.network.neutron [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.715219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.715685] env[61991]: DEBUG nova.objects.instance [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'migration_context' on Instance uuid d36bd981-9ea3-46f6-8376-ac1e0c3bf61e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.806111] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130294, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.928572] env[61991]: DEBUG oslo_vmware.api [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130293, 'name': PowerOnVM_Task, 'duration_secs': 1.40001} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.928921] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.929181] env[61991]: INFO nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Took 9.25 seconds to spawn the instance on the hypervisor. [ 1006.929375] env[61991]: DEBUG nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.930388] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6126b9c3-5f20-4098-8dd7-691e80f121ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.958651] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.218343] env[61991]: DEBUG nova.objects.base [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1007.219327] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c311721-9507-4fc3-8678-e0b370899844 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.237984] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aabe1c48-d66a-4311-b0a2-2e4affc9dd67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.242906] env[61991]: DEBUG oslo_vmware.api [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1007.242906] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284d8f3-8c5a-6e51-5a39-c85a937a8815" [ 1007.242906] env[61991]: _type = "Task" [ 1007.242906] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.250035] env[61991]: DEBUG oslo_vmware.api [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284d8f3-8c5a-6e51-5a39-c85a937a8815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.305308] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130294, 'name': ReconfigVM_Task, 'duration_secs': 1.16365} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.305614] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1007.306307] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79c74ab-16c8-4678-a58f-84d87d77980b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.327185] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d/8005b8f6-6f4d-4f72-be5d-955fa114408d.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.327473] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6749c5cb-fa57-4b24-a363-6f54301b5690 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.349281] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1007.349281] env[61991]: value = "task-1130296" [ 1007.349281] env[61991]: _type = "Task" [ 1007.349281] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.357408] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130296, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.450406] env[61991]: INFO nova.compute.manager [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Took 17.61 seconds to build instance. [ 1007.471857] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ada6322-e114-4c1a-a526-9e0dfac94e61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.480379] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd441ea-d4c7-4002-a4d4-a5f824b71829 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.755235] env[61991]: DEBUG oslo_vmware.api [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5284d8f3-8c5a-6e51-5a39-c85a937a8815, 'name': SearchDatastore_Task, 'duration_secs': 0.00608} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.755492] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.755724] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.859716] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130296, 'name': ReconfigVM_Task, 'duration_secs': 0.247296} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.859949] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d/8005b8f6-6f4d-4f72-be5d-955fa114408d.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.860234] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1007.952328] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2c943e2b-89e0-40a0-bfa5-ae8be8d8eec1 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.127s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.049201] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.049543] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.049822] env[61991]: INFO nova.compute.manager [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Rebooting instance [ 1008.368372] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2837e0d-eb30-426d-8026-af1cbdb4c9f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.388878] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02152ef9-9d58-41a9-b289-8647bad81288 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.407483] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.441163] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b73cc1-8092-43d1-88b0-ae70805ac46f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.448044] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af5d2b0-fe52-4def-94f5-1733571220b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.479408] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5572c8bc-a450-4144-9dd1-3551bd0f3d69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.486303] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ded4c1e-9835-4673-9445-3a81313fd614 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.500161] env[61991]: DEBUG nova.compute.provider_tree [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.566725] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.566911] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquired lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.567106] env[61991]: DEBUG nova.network.neutron [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1008.585141] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48895d22-a7c1-4fbc-b453-36d0ed8fb9d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.607319] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae9e647-365e-40ec-b664-60134cc126ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.614529] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.957932] env[61991]: DEBUG nova.network.neutron [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Port 15cc98a6-0818-4630-b737-e50e1fb71851 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1009.019962] env[61991]: ERROR nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [req-226bcb7d-bb73-4047-8aec-22bddadc82e4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID d748992a-e0bf-4ec2-9c17-0e373360e5a3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-226bcb7d-bb73-4047-8aec-22bddadc82e4"}]} [ 1009.035241] env[61991]: DEBUG nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1009.047633] env[61991]: DEBUG nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1009.047844] env[61991]: DEBUG nova.compute.provider_tree [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 167, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.057866] env[61991]: DEBUG nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1009.077724] env[61991]: DEBUG nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1009.121532] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1009.121946] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c27f0d29-1ab5-459f-9cf2-5db5292a1df9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.131230] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1009.131230] env[61991]: value = "task-1130298" [ 1009.131230] env[61991]: _type = "Task" [ 1009.131230] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.139303] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130298, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.256295] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd0a849-2461-4495-9fe3-4b18a8b4a80d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.264631] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c38f9c0-6535-47fa-9466-dffd2f06f6bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.299596] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca56c9b3-e469-4033-bf01-fff22e8f991f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.308059] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7414abaa-e629-4968-ae75-7b46d8703d84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.322588] env[61991]: DEBUG nova.compute.provider_tree [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.337645] env[61991]: DEBUG nova.network.neutron [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Updating instance_info_cache with network_info: [{"id": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "address": "fa:16:3e:50:57:72", "network": {"id": "ca86d109-d0dd-490f-be31-686b0fed7f9e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-2065252436-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "860b2205deb34624b529effcc1f0b44f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae215ba8-f7a5-4b23-a055-90316d29817f", "external-id": "nsx-vlan-transportzone-798", "segmentation_id": 798, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16df29ee-00", "ovs_interfaceid": "16df29ee-00a7-4144-8e1b-3b0c0699817e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.642401] env[61991]: DEBUG oslo_vmware.api [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130298, 'name': PowerOnVM_Task, 'duration_secs': 0.411521} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.642712] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.642858] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1bb6e560-83df-4f2a-850b-dc3cf4fa4d03 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance '1730f9b3-69ad-476b-b3ab-3e1345f2a115' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1009.840457] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Releasing lock "refresh_cache-39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.843292] env[61991]: DEBUG nova.compute.manager [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.844513] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2811ac-0a06-420b-a53f-3ab42436f7bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.855475] env[61991]: DEBUG nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updated inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1009.855737] env[61991]: DEBUG nova.compute.provider_tree [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updating resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 generation from 149 to 150 during operation: update_inventory {{(pid=61991) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1009.855924] env[61991]: DEBUG nova.compute.provider_tree [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.981399] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.981653] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.982029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.869188] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.113s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.875125] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5be4fcf-bf57-48e1-8b2a-040deff977ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.883978] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Doing hard reboot of VM {{(pid=61991) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1010.884296] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-a5b05af7-44b9-42d0-9a8f-ad51d4250305 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.891824] env[61991]: DEBUG oslo_vmware.api [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1010.891824] env[61991]: value = "task-1130300" [ 1010.891824] env[61991]: _type = "Task" [ 1010.891824] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.901534] env[61991]: DEBUG oslo_vmware.api [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130300, 'name': ResetVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.022023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.022248] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.022432] env[61991]: DEBUG nova.network.neutron [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.403906] env[61991]: DEBUG oslo_vmware.api [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130300, 'name': ResetVM_Task, 'duration_secs': 0.106077} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.404206] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Did hard reboot of VM {{(pid=61991) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1011.404408] env[61991]: DEBUG nova.compute.manager [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.405203] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c40747e-529a-48c1-b787-40aa3e1a9334 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.432641] env[61991]: INFO nova.scheduler.client.report [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted allocation for migration ae50fc19-9498-4e5f-9b4f-c47537530c53 [ 1011.571275] env[61991]: DEBUG nova.network.neutron [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Port 4e506178-06a6-49c6-a6b3-fa24ebafc988 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1011.571559] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.571715] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.571880] env[61991]: DEBUG nova.network.neutron [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.739989] env[61991]: DEBUG nova.network.neutron [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance_info_cache with network_info: [{"id": "15cc98a6-0818-4630-b737-e50e1fb71851", "address": "fa:16:3e:97:a6:51", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15cc98a6-08", "ovs_interfaceid": "15cc98a6-0818-4630-b737-e50e1fb71851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.918037] env[61991]: DEBUG oslo_concurrency.lockutils [None req-5e713c70-f2b8-4b09-8ab5-1982d6a2f22d tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.868s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.937521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-62a8c04f-e5c7-4f61-a9f1-e83371f5cfed tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.473s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.242784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.368119] env[61991]: DEBUG nova.network.neutron [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.771125] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce1c7e9-468e-4799-9a3c-d62f1bb4c4f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.791152] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b06d3b9-db5e-4c41-8d54-e30eaae1dfc6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.798487] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1012.848751] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.849047] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.849274] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.849464] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.849643] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.851721] env[61991]: INFO nova.compute.manager [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Terminating instance [ 1012.853888] env[61991]: DEBUG nova.compute.manager [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1012.854099] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.854992] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4154cc6-9358-482c-bd3f-a02b70ea63a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.857916] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.858138] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.858327] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.858509] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.858676] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.860537] env[61991]: INFO nova.compute.manager [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Terminating instance [ 1012.862290] env[61991]: DEBUG nova.compute.manager [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1012.862481] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.863290] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f72971b-504f-4ed8-969e-20710b4c2a52 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.868301] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.868890] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-680bd943-8dec-459c-bafa-aa694933dbce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.870646] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.875588] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.876044] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a204936c-ab45-4fcc-b4ce-a99656f4280d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.881965] env[61991]: DEBUG oslo_vmware.api [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1012.881965] env[61991]: value = "task-1130301" [ 1012.881965] env[61991]: _type = "Task" [ 1012.881965] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.889412] env[61991]: DEBUG oslo_vmware.api [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1012.889412] env[61991]: value = "task-1130302" [ 1012.889412] env[61991]: _type = "Task" [ 1012.889412] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.893421] env[61991]: DEBUG oslo_vmware.api [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.906344] env[61991]: DEBUG oslo_vmware.api [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130302, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.305220] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.305544] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e33b7ee3-803c-4333-ab40-a6f26e5a6e49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.313875] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1013.313875] env[61991]: value = "task-1130303" [ 1013.313875] env[61991]: _type = "Task" [ 1013.313875] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.323745] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130303, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.377570] env[61991]: DEBUG nova.compute.manager [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61991) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1013.395116] env[61991]: DEBUG oslo_vmware.api [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130301, 'name': PowerOffVM_Task, 'duration_secs': 0.207204} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.398401] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.398627] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.398842] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8a722b0-612a-4abc-94c3-60be6163d808 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.407196] env[61991]: DEBUG oslo_vmware.api [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130302, 'name': PowerOffVM_Task, 'duration_secs': 0.221645} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.407530] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.407716] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.407965] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-952c2742-5e4f-4141-8ff7-9935f4db57b8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.486731] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.486993] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.487221] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Deleting the datastore file [datastore2] 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.487499] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbdd2a54-e119-4de7-858c-733fe9b4c0f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.491855] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.492784] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.492784] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleting the datastore file [datastore2] d36bd981-9ea3-46f6-8376-ac1e0c3bf61e {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.492784] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-abfafbe7-1143-4cdc-909f-fc1a662f3abd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.499206] env[61991]: DEBUG oslo_vmware.api [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for the task: (returnval){ [ 1013.499206] env[61991]: value = "task-1130306" [ 1013.499206] env[61991]: _type = "Task" [ 1013.499206] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.500696] env[61991]: DEBUG oslo_vmware.api [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1013.500696] env[61991]: value = "task-1130307" [ 1013.500696] env[61991]: _type = "Task" [ 1013.500696] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.511939] env[61991]: DEBUG oslo_vmware.api [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.514907] env[61991]: DEBUG oslo_vmware.api [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.824698] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130303, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.011184] env[61991]: DEBUG oslo_vmware.api [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Task: {'id': task-1130306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145111} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.011894] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.012164] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.012379] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.012562] env[61991]: INFO nova.compute.manager [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1014.012803] env[61991]: DEBUG oslo.service.loopingcall [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.013292] env[61991]: DEBUG nova.compute.manager [-] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.013393] env[61991]: DEBUG nova.network.neutron [-] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.018281] env[61991]: DEBUG oslo_vmware.api [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158954} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.018809] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.018993] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.019190] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.019364] env[61991]: INFO nova.compute.manager [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1014.019589] env[61991]: DEBUG oslo.service.loopingcall [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.019773] env[61991]: DEBUG nova.compute.manager [-] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.019863] env[61991]: DEBUG nova.network.neutron [-] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1014.300874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.301167] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.304108] env[61991]: DEBUG nova.compute.manager [req-a111263f-e32e-436a-9630-40bd3b6ec47e req-e8a91924-052a-4405-9921-3cb9d0a69dad service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Received event network-vif-deleted-16df29ee-00a7-4144-8e1b-3b0c0699817e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.304108] env[61991]: INFO nova.compute.manager [req-a111263f-e32e-436a-9630-40bd3b6ec47e req-e8a91924-052a-4405-9921-3cb9d0a69dad service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Neutron deleted interface 16df29ee-00a7-4144-8e1b-3b0c0699817e; detaching it from the instance and deleting it from the info cache [ 1014.304108] env[61991]: DEBUG nova.network.neutron [req-a111263f-e32e-436a-9630-40bd3b6ec47e req-e8a91924-052a-4405-9921-3cb9d0a69dad service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.325044] env[61991]: DEBUG oslo_vmware.api [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130303, 'name': PowerOnVM_Task, 'duration_secs': 0.573478} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.325044] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.325498] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5310bb-85bf-4313-be76-95e0b552885c tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance '8005b8f6-6f4d-4f72-be5d-955fa114408d' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.472961] env[61991]: DEBUG nova.compute.manager [req-38220f53-e606-41ef-a0ec-bfbef8e63ec0 req-ded20ff7-59d5-4ef5-a1e2-8a7d14f25327 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Received event network-vif-deleted-9452191d-e715-4ad9-a48e-43aebb90b551 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.473139] env[61991]: INFO nova.compute.manager [req-38220f53-e606-41ef-a0ec-bfbef8e63ec0 req-ded20ff7-59d5-4ef5-a1e2-8a7d14f25327 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Neutron deleted interface 9452191d-e715-4ad9-a48e-43aebb90b551; detaching it from the instance and deleting it from the info cache [ 1014.473322] env[61991]: DEBUG nova.network.neutron [req-38220f53-e606-41ef-a0ec-bfbef8e63ec0 req-ded20ff7-59d5-4ef5-a1e2-8a7d14f25327 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.485118] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.485411] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.777325] env[61991]: DEBUG nova.network.neutron [-] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.805524] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1014.808441] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18a7d3e7-7aa1-4b5a-91b6-bb8b19d7e816 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.819732] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d9c9af-adc0-438c-af22-990ddde47705 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.857088] env[61991]: DEBUG nova.compute.manager [req-a111263f-e32e-436a-9630-40bd3b6ec47e req-e8a91924-052a-4405-9921-3cb9d0a69dad service nova] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Detach interface failed, port_id=16df29ee-00a7-4144-8e1b-3b0c0699817e, reason: Instance 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1014.947579] env[61991]: DEBUG nova.network.neutron [-] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.977558] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2383105c-45a2-44b6-9bb0-c9536dea5244 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.987024] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cab79a-dec2-4cf2-bf72-ec1907c2ca8f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.997761] env[61991]: DEBUG nova.objects.instance [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'migration_context' on Instance uuid 1730f9b3-69ad-476b-b3ab-3e1345f2a115 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.020320] env[61991]: DEBUG nova.compute.manager [req-38220f53-e606-41ef-a0ec-bfbef8e63ec0 req-ded20ff7-59d5-4ef5-a1e2-8a7d14f25327 service nova] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Detach interface failed, port_id=9452191d-e715-4ad9-a48e-43aebb90b551, reason: Instance d36bd981-9ea3-46f6-8376-ac1e0c3bf61e could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1015.280566] env[61991]: INFO nova.compute.manager [-] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Took 1.27 seconds to deallocate network for instance. [ 1015.332922] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.450462] env[61991]: INFO nova.compute.manager [-] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Took 1.43 seconds to deallocate network for instance. [ 1015.663721] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3e554e-3878-46d6-b7f8-fdbb1d22d1af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.672679] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459eec23-3be1-4488-849a-ff1b0d3ef5b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.703133] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276f570c-8cdd-46dd-bb68-9ec107dfcb5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.711460] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bd1b6f-1259-4697-b2a4-0a85691b3cc5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.725100] env[61991]: DEBUG nova.compute.provider_tree [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.787406] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.958315] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.227793] env[61991]: DEBUG nova.scheduler.client.report [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.609363] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.609719] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.609763] env[61991]: DEBUG nova.compute.manager [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Going to confirm migration 7 {{(pid=61991) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1016.832795] env[61991]: DEBUG oslo_concurrency.lockutils [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.833090] env[61991]: DEBUG oslo_concurrency.lockutils [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.155856] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.156076] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.156265] env[61991]: DEBUG nova.network.neutron [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1017.156496] env[61991]: DEBUG nova.objects.instance [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lazy-loading 'info_cache' on Instance uuid 8005b8f6-6f4d-4f72-be5d-955fa114408d {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.239398] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.754s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.245482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.913s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.247829] env[61991]: INFO nova.compute.claims [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.335844] env[61991]: INFO nova.compute.manager [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Detaching volume 3d090d84-cb6e-4050-b26a-89d77c479ef4 [ 1017.368890] env[61991]: INFO nova.virt.block_device [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Attempting to driver detach volume 3d090d84-cb6e-4050-b26a-89d77c479ef4 from mountpoint /dev/sdb [ 1017.369241] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1017.369570] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246944', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'name': 'volume-3d090d84-cb6e-4050-b26a-89d77c479ef4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5364f93c-ca53-480b-8397-c331dfdeed3f', 'attached_at': '', 'detached_at': '', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'serial': '3d090d84-cb6e-4050-b26a-89d77c479ef4'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1017.370940] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76dc3c09-8e96-40c7-bc9f-e0d000b9eb84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.403560] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4220ba6-ebf1-4de8-bd3f-ddcf0e2f76d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.414111] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790f5a13-8fb5-410d-b13d-acc52b834ccc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.443588] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4eafbd7-4ab6-4a88-ac43-e2ba97a95b99 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.461665] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] The volume has not been displaced from its original location: [datastore1] volume-3d090d84-cb6e-4050-b26a-89d77c479ef4/volume-3d090d84-cb6e-4050-b26a-89d77c479ef4.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1017.466813] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfiguring VM instance instance-0000005c to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1017.467208] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-466d540a-2e4a-4f58-bb34-50ac0c9b9687 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.485956] env[61991]: DEBUG oslo_vmware.api [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1017.485956] env[61991]: value = "task-1130308" [ 1017.485956] env[61991]: _type = "Task" [ 1017.485956] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.495298] env[61991]: DEBUG oslo_vmware.api [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130308, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.996060] env[61991]: DEBUG oslo_vmware.api [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130308, 'name': ReconfigVM_Task, 'duration_secs': 0.262719} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.996377] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Reconfigured VM instance instance-0000005c to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1018.000862] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b47b607-d034-40a1-a837-6c28def7f877 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.016219] env[61991]: DEBUG oslo_vmware.api [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1018.016219] env[61991]: value = "task-1130309" [ 1018.016219] env[61991]: _type = "Task" [ 1018.016219] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.025887] env[61991]: DEBUG oslo_vmware.api [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130309, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.365290] env[61991]: DEBUG nova.network.neutron [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance_info_cache with network_info: [{"id": "15cc98a6-0818-4630-b737-e50e1fb71851", "address": "fa:16:3e:97:a6:51", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15cc98a6-08", "ovs_interfaceid": "15cc98a6-0818-4630-b737-e50e1fb71851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.434223] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9600ebe8-85f3-440d-8831-a2858f95d6a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.442878] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19414cb1-a01b-4f21-a92e-e90e239213cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.473849] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56b6400-eddf-4ecf-9eff-ebe596f5091c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.481803] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751bdf47-7b0f-45a4-a11f-05feb3af8e34 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.495177] env[61991]: DEBUG nova.compute.provider_tree [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.526214] env[61991]: DEBUG oslo_vmware.api [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130309, 'name': ReconfigVM_Task, 'duration_secs': 0.16212} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.526567] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246944', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'name': 'volume-3d090d84-cb6e-4050-b26a-89d77c479ef4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5364f93c-ca53-480b-8397-c331dfdeed3f', 'attached_at': '', 'detached_at': '', 'volume_id': '3d090d84-cb6e-4050-b26a-89d77c479ef4', 'serial': '3d090d84-cb6e-4050-b26a-89d77c479ef4'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1018.787701] env[61991]: INFO nova.compute.manager [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Swapping old allocation on dict_keys(['d748992a-e0bf-4ec2-9c17-0e373360e5a3']) held by migration 09b55899-bef6-4618-9ecd-0c7519b9cac5 for instance [ 1018.809211] env[61991]: DEBUG nova.scheduler.client.report [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Overwriting current allocation {'allocations': {'d748992a-e0bf-4ec2-9c17-0e373360e5a3': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 152}}, 'project_id': '822d7e3c678e4defa24bb4d8439a62de', 'user_id': '0c5daf21b21b40c9a54bed64a64538bb', 'consumer_generation': 1} on consumer 1730f9b3-69ad-476b-b3ab-3e1345f2a115 {{(pid=61991) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1018.869487] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-8005b8f6-6f4d-4f72-be5d-955fa114408d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.869760] env[61991]: DEBUG nova.objects.instance [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lazy-loading 'migration_context' on Instance uuid 8005b8f6-6f4d-4f72-be5d-955fa114408d {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.887049] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.887255] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.887436] env[61991]: DEBUG nova.network.neutron [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.998964] env[61991]: DEBUG nova.scheduler.client.report [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.072157] env[61991]: DEBUG nova.objects.instance [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'flavor' on Instance uuid 5364f93c-ca53-480b-8397-c331dfdeed3f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.372269] env[61991]: DEBUG nova.objects.base [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Object Instance<8005b8f6-6f4d-4f72-be5d-955fa114408d> lazy-loaded attributes: info_cache,migration_context {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1019.373287] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9d6ddf-f3ff-415a-98a6-e821284ccc8b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.395131] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7299279-c426-40af-803c-dcccd6dd3940 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.401043] env[61991]: DEBUG oslo_vmware.api [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1019.401043] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52324e13-27f3-2cda-0a4d-486b5ddcd878" [ 1019.401043] env[61991]: _type = "Task" [ 1019.401043] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.408184] env[61991]: DEBUG oslo_vmware.api [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52324e13-27f3-2cda-0a4d-486b5ddcd878, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.504353] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.504667] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1019.507408] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.720s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.507658] env[61991]: DEBUG nova.objects.instance [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lazy-loading 'resources' on Instance uuid 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.608323] env[61991]: DEBUG nova.network.neutron [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [{"id": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "address": "fa:16:3e:e0:7e:40", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e506178-06", "ovs_interfaceid": "4e506178-06a6-49c6-a6b3-fa24ebafc988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.912430] env[61991]: DEBUG oslo_vmware.api [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52324e13-27f3-2cda-0a4d-486b5ddcd878, 'name': SearchDatastore_Task, 'duration_secs': 0.008258} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.912726] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.011594] env[61991]: DEBUG nova.compute.utils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.016866] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1020.017112] env[61991]: DEBUG nova.network.neutron [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.059011] env[61991]: DEBUG nova.policy [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b5d90a913fc41e282e050bb9af628d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adbd4ea9b0744f28a0d4a46b4a04d683', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.078955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-657d75be-23d9-4442-bd0b-d08941df1e46 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.246s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.111339] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-1730f9b3-69ad-476b-b3ab-3e1345f2a115" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.112566] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8643a91f-e014-458d-b405-e2fb8c20ef8e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.122546] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffde6f6d-ba38-43c5-b806-6ff4f2f654f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.171676] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9142e6c0-f438-4a97-8a43-89566aad741a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.179720] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a86b67-5e9e-4405-a700-b04e04c98028 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.210346] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19df58d-fff5-4788-b806-14a17f385d6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.218782] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394cc3fb-e1f4-4d09-9efd-0c6d42f9cf54 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.233378] env[61991]: DEBUG nova.compute.provider_tree [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.418298] env[61991]: DEBUG nova.network.neutron [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Successfully created port: 7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.518153] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1020.736687] env[61991]: DEBUG nova.scheduler.client.report [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.025588] env[61991]: INFO nova.virt.block_device [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Booting with volume f529fecd-03fc-4258-a8dc-f770503a03ee at /dev/sda [ 1021.060156] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac9ed86c-f361-48f6-a36e-1bb1169c7289 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.070500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533f66a4-2d91-4c0b-8152-8a62cd9d826a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.104915] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff1d7f27-5c20-4433-9068-9f11e1d53be9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.114435] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f163c6-5858-49dd-aef8-394a78f8f58a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.145379] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a726ec-eb2b-484d-92ca-f92cdf76f37f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.152157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d80da8-4774-4f9e-8330-0c587b2c7526 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.165652] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.165911] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.166145] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "5364f93c-ca53-480b-8397-c331dfdeed3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.166335] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.166539] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.168193] env[61991]: DEBUG nova.virt.block_device [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating existing volume attachment record: 5625e464-19bb-4cfc-9019-33372acf3876 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1021.170390] env[61991]: INFO nova.compute.manager [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Terminating instance [ 1021.172170] env[61991]: DEBUG nova.compute.manager [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.172364] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.173086] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cd9350-06ad-4409-9053-099d377f6921 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.181743] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.181972] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f785f894-c296-473f-a04e-e9686f8beecc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.188673] env[61991]: DEBUG oslo_vmware.api [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1021.188673] env[61991]: value = "task-1130310" [ 1021.188673] env[61991]: _type = "Task" [ 1021.188673] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.197123] env[61991]: DEBUG oslo_vmware.api [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.225948] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.226291] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0260b0dc-1e34-4047-83b6-0a503ff87980 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.234718] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1021.234718] env[61991]: value = "task-1130311" [ 1021.234718] env[61991]: _type = "Task" [ 1021.234718] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.241672] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.250028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.291s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.250028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.252138] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.339s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.253779] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130311, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.272146] env[61991]: INFO nova.scheduler.client.report [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted allocations for instance d36bd981-9ea3-46f6-8376-ac1e0c3bf61e [ 1021.274079] env[61991]: INFO nova.scheduler.client.report [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Deleted allocations for instance 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6 [ 1021.698773] env[61991]: DEBUG oslo_vmware.api [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130310, 'name': PowerOffVM_Task, 'duration_secs': 0.18526} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.699045] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.699236] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.699489] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-363bd6c1-291a-47a9-b0d5-ecf32e9dc46c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.745687] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130311, 'name': PowerOffVM_Task, 'duration_secs': 0.24543} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.746039] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.746831] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.747128] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.747327] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.747576] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.747735] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.748066] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.748119] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.748304] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.748467] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.748634] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.748815] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.753810] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56d54d45-6b46-433a-93ad-5654ceff372b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.769449] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.769755] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.771990] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleting the datastore file [datastore2] 5364f93c-ca53-480b-8397-c331dfdeed3f {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.771990] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0682de61-9b85-4fb4-a131-c1397b80b5c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.773843] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1021.773843] env[61991]: value = "task-1130313" [ 1021.773843] env[61991]: _type = "Task" [ 1021.773843] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.778144] env[61991]: DEBUG oslo_vmware.api [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1021.778144] env[61991]: value = "task-1130314" [ 1021.778144] env[61991]: _type = "Task" [ 1021.778144] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.789375] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130313, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.789875] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9a71f9da-24bc-4d7f-991e-59d3dbf425e7 tempest-InstanceActionsTestJSON-1275513061 tempest-InstanceActionsTestJSON-1275513061-project-member] Lock "39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.932s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.793710] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9359f12c-e771-4cad-a53c-9bcefeee4ff6 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "d36bd981-9ea3-46f6-8376-ac1e0c3bf61e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.944s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.800890] env[61991]: DEBUG oslo_vmware.api [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.960802] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429d9a79-2a42-4f7c-b100-577365b18846 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.969949] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fe790e-a07d-424e-8e26-92bd0ed0b4c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.974115] env[61991]: DEBUG nova.compute.manager [req-9858fc81-cfb4-4c5f-99e6-49d4c15b64a6 req-bfe14d8e-a0c2-412c-9f9c-6b1464b696ee service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Received event network-vif-plugged-7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.974327] env[61991]: DEBUG oslo_concurrency.lockutils [req-9858fc81-cfb4-4c5f-99e6-49d4c15b64a6 req-bfe14d8e-a0c2-412c-9f9c-6b1464b696ee service nova] Acquiring lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.974567] env[61991]: DEBUG oslo_concurrency.lockutils [req-9858fc81-cfb4-4c5f-99e6-49d4c15b64a6 req-bfe14d8e-a0c2-412c-9f9c-6b1464b696ee service nova] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.974741] env[61991]: DEBUG oslo_concurrency.lockutils [req-9858fc81-cfb4-4c5f-99e6-49d4c15b64a6 req-bfe14d8e-a0c2-412c-9f9c-6b1464b696ee service nova] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.974909] env[61991]: DEBUG nova.compute.manager [req-9858fc81-cfb4-4c5f-99e6-49d4c15b64a6 req-bfe14d8e-a0c2-412c-9f9c-6b1464b696ee service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] No waiting events found dispatching network-vif-plugged-7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.975200] env[61991]: WARNING nova.compute.manager [req-9858fc81-cfb4-4c5f-99e6-49d4c15b64a6 req-bfe14d8e-a0c2-412c-9f9c-6b1464b696ee service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Received unexpected event network-vif-plugged-7277dc2e-f381-41e1-95de-31eda0f6cb27 for instance with vm_state building and task_state block_device_mapping. [ 1022.004486] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34328f7a-bf64-438d-98c4-048af3e4998a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.013229] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8881090-a655-433d-b433-0fa31abc1e7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.026606] env[61991]: DEBUG nova.compute.provider_tree [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.287728] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130313, 'name': ReconfigVM_Task, 'duration_secs': 0.171064} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.291935] env[61991]: DEBUG oslo_vmware.api [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157607} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.292668] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b749223-a966-4884-bd72-5469ca7bec0e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.295339] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.295538] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.295722] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.295901] env[61991]: INFO nova.compute.manager [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1022.296154] env[61991]: DEBUG oslo.service.loopingcall [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.296995] env[61991]: DEBUG nova.compute.manager [-] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.297110] env[61991]: DEBUG nova.network.neutron [-] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1022.317441] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.317707] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.317872] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.318073] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.318229] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.318380] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.318588] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.318756] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.318938] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.319115] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.319296] env[61991]: DEBUG nova.virt.hardware [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.320094] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66e33181-084e-41d9-b63f-fb6447fb540a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.327047] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1022.327047] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525427f2-ac2a-6aaa-e8ef-e81193988a5c" [ 1022.327047] env[61991]: _type = "Task" [ 1022.327047] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.336083] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525427f2-ac2a-6aaa-e8ef-e81193988a5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.532381] env[61991]: DEBUG nova.scheduler.client.report [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.542249] env[61991]: DEBUG nova.network.neutron [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Successfully updated port: 7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.700804] env[61991]: DEBUG nova.compute.manager [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Received event network-changed-7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.701034] env[61991]: DEBUG nova.compute.manager [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Refreshing instance network info cache due to event network-changed-7277dc2e-f381-41e1-95de-31eda0f6cb27. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1022.701270] env[61991]: DEBUG oslo_concurrency.lockutils [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] Acquiring lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.701418] env[61991]: DEBUG oslo_concurrency.lockutils [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] Acquired lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.701583] env[61991]: DEBUG nova.network.neutron [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Refreshing network info cache for port 7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.558905] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.561824] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1023.562310] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.562525] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.562717] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.562927] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.563087] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.563230] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.563425] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.563585] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.563785] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.563952] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.564160] env[61991]: DEBUG nova.virt.hardware [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.566063] env[61991]: DEBUG nova.compute.manager [req-9f472c9f-dd41-4532-a736-da3e2e14f8b1 req-2fb58700-baa1-4a1f-b4d8-217f311fab78 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Received event network-vif-deleted-2fc2099f-0ea4-4665-948a-01d944f9275a {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.566063] env[61991]: INFO nova.compute.manager [req-9f472c9f-dd41-4532-a736-da3e2e14f8b1 req-2fb58700-baa1-4a1f-b4d8-217f311fab78 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Neutron deleted interface 2fc2099f-0ea4-4665-948a-01d944f9275a; detaching it from the instance and deleting it from the info cache [ 1023.566063] env[61991]: DEBUG nova.network.neutron [req-9f472c9f-dd41-4532-a736-da3e2e14f8b1 req-2fb58700-baa1-4a1f-b4d8-217f311fab78 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.567822] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503b62d2-5659-478e-b542-52c124092640 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.577657] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525427f2-ac2a-6aaa-e8ef-e81193988a5c, 'name': SearchDatastore_Task, 'duration_secs': 0.014319} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.584372] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.585525] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcaf0c6c-428b-4873-9d04-f694c155e03e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.599037] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc39627-59aa-4608-9797-831d8217e1b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.617280] env[61991]: DEBUG nova.network.neutron [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.619782] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1023.619782] env[61991]: value = "task-1130315" [ 1023.619782] env[61991]: _type = "Task" [ 1023.619782] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.630219] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.646050] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.647112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.797676] env[61991]: DEBUG nova.network.neutron [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.957180] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1023.957413] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1023.957581] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1023.957708] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1024.068320] env[61991]: DEBUG nova.network.neutron [-] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.071115] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.819s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.074305] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3854a46-4b43-4f50-941b-6f0fcc373e1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.086416] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec3e6c4-ac8c-427b-812b-a68c60a7f730 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.125073] env[61991]: DEBUG nova.compute.manager [req-9f472c9f-dd41-4532-a736-da3e2e14f8b1 req-2fb58700-baa1-4a1f-b4d8-217f311fab78 service nova] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Detach interface failed, port_id=2fc2099f-0ea4-4665-948a-01d944f9275a, reason: Instance 5364f93c-ca53-480b-8397-c331dfdeed3f could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.135019] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130315, 'name': ReconfigVM_Task, 'duration_secs': 0.312279} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.135266] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1024.136136] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d198fde-e358-4e47-a2aa-a9fefbd2a6a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.155943] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1024.166993] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.167826] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c49c219-e3ec-4752-959d-03f459de85ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.189545] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1024.189545] env[61991]: value = "task-1130316" [ 1024.189545] env[61991]: _type = "Task" [ 1024.189545] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.199075] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130316, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.300425] env[61991]: DEBUG oslo_concurrency.lockutils [req-4fad586a-8ef2-45c5-ba10-e5c84a30c32c req-d780ff10-701f-4ca0-9afd-cecfe73c9bb8 service nova] Releasing lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.300858] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.301051] env[61991]: DEBUG nova.network.neutron [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.462279] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Skipping network cache update for instance because it is being deleted. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1024.462476] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1024.504268] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.504518] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.504682] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1024.504835] env[61991]: DEBUG nova.objects.instance [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lazy-loading 'info_cache' on Instance uuid cf99c9cc-24c3-4acc-8120-49c4b12a3553 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.574924] env[61991]: INFO nova.compute.manager [-] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Took 2.28 seconds to deallocate network for instance. [ 1024.651405] env[61991]: INFO nova.scheduler.client.report [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocation for migration 5a1d3705-160c-468f-a272-eff9cc47f565 [ 1024.689698] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.689975] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.691645] env[61991]: INFO nova.compute.claims [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.704672] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130316, 'name': ReconfigVM_Task, 'duration_secs': 0.276715} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.704985] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115/1730f9b3-69ad-476b-b3ab-3e1345f2a115.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.705859] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd097435-c074-4391-96d7-fa4955d6c69b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.728789] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b6d36f-b78b-4c33-8b0f-5ec2298d8ee6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.752201] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94c4c4e-f695-4448-9899-c5354ec85e8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.757974] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "892323a3-8661-47da-a52a-d7e0557ea300" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.758323] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "892323a3-8661-47da-a52a-d7e0557ea300" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.758498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "892323a3-8661-47da-a52a-d7e0557ea300-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.758646] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "892323a3-8661-47da-a52a-d7e0557ea300-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.758829] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "892323a3-8661-47da-a52a-d7e0557ea300-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.776890] env[61991]: INFO nova.compute.manager [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Terminating instance [ 1024.779416] env[61991]: DEBUG nova.compute.manager [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.779622] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.780441] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9921be-529d-48a1-be1e-7cc6c064a81a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.784273] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f748206-6fa0-489c-98c0-414aad8f7f31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.792115] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1024.794417] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f2b08d1-d7b0-4125-95bd-8f2af0251fd3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.796110] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.796352] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7486794d-f7e3-4021-9387-43b2c275ba33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.806240] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1024.806240] env[61991]: value = "task-1130317" [ 1024.806240] env[61991]: _type = "Task" [ 1024.806240] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.807641] env[61991]: DEBUG oslo_vmware.api [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 1024.807641] env[61991]: value = "task-1130318" [ 1024.807641] env[61991]: _type = "Task" [ 1024.807641] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.819525] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.822736] env[61991]: DEBUG oslo_vmware.api [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130318, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.848846] env[61991]: DEBUG nova.network.neutron [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1025.083078] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.098514] env[61991]: DEBUG nova.network.neutron [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [{"id": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "address": "fa:16:3e:37:4b:7d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7277dc2e-f3", "ovs_interfaceid": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.116654] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.156497] env[61991]: DEBUG oslo_concurrency.lockutils [None req-eff2413e-3869-4c48-9b63-5aa33d2d1dcf tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.547s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.157764] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.041s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.158114] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.158453] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.158859] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.161213] env[61991]: INFO nova.compute.manager [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Terminating instance [ 1025.163370] env[61991]: DEBUG nova.compute.manager [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.163568] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.164668] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e00fd5d-2c70-4f47-8b48-e86b5c0634f3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.173658] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.174529] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-669f0d17-18cd-47fb-aab9-28fddf45cf8b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.185022] env[61991]: DEBUG oslo_vmware.api [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1025.185022] env[61991]: value = "task-1130319" [ 1025.185022] env[61991]: _type = "Task" [ 1025.185022] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.191275] env[61991]: DEBUG oslo_vmware.api [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.321788] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130317, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.325102] env[61991]: DEBUG oslo_vmware.api [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130318, 'name': PowerOffVM_Task, 'duration_secs': 0.204059} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.325452] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.325703] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.326043] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6245784-a796-4d1b-974f-fc097aaefa2e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.426772] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.427066] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.427761] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Deleting the datastore file [datastore1] 892323a3-8661-47da-a52a-d7e0557ea300 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.427761] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84c5d092-c5bc-4f22-8667-c85ee007330f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.435683] env[61991]: DEBUG oslo_vmware.api [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for the task: (returnval){ [ 1025.435683] env[61991]: value = "task-1130321" [ 1025.435683] env[61991]: _type = "Task" [ 1025.435683] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.445995] env[61991]: DEBUG oslo_vmware.api [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130321, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.600801] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.601237] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Instance network_info: |[{"id": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "address": "fa:16:3e:37:4b:7d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7277dc2e-f3", "ovs_interfaceid": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.601616] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:4b:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dced2f3d-7fd3-4a42-836d-9f02dab4c949', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7277dc2e-f381-41e1-95de-31eda0f6cb27', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.610625] env[61991]: DEBUG oslo.service.loopingcall [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.611399] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.612303] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-992338ef-0a1b-40eb-a037-ec6da6fc2779 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.630560] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "81736f13-fc9e-4246-9d86-5a5bd7e15832" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.630680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.635954] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.635954] env[61991]: value = "task-1130322" [ 1025.635954] env[61991]: _type = "Task" [ 1025.635954] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.645495] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130322, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.694741] env[61991]: DEBUG oslo_vmware.api [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130319, 'name': PowerOffVM_Task, 'duration_secs': 0.314464} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.694741] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.694741] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.694741] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de1c57ab-4677-42b0-b397-cc00e90a7b89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.773429] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.773691] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.773878] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore1] 8005b8f6-6f4d-4f72-be5d-955fa114408d {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.774224] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d8dd707-c083-4f10-a1ca-928dd630e870 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.786381] env[61991]: DEBUG oslo_vmware.api [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1025.786381] env[61991]: value = "task-1130324" [ 1025.786381] env[61991]: _type = "Task" [ 1025.786381] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.797993] env[61991]: DEBUG oslo_vmware.api [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.818437] env[61991]: DEBUG oslo_vmware.api [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130317, 'name': PowerOnVM_Task, 'duration_secs': 0.572394} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.818437] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1025.891992] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f78fb83-2878-423b-b053-f6d49bdff287 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.902035] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67fab1e-6474-4526-9163-e7f4355ae2b3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.932698] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb79ea9-19d9-4d12-a343-05acf740d549 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.943186] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103bad55-bbf3-4d1d-b116-a1a921725c35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.950673] env[61991]: DEBUG oslo_vmware.api [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Task: {'id': task-1130321, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151803} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.958156] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.958382] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.958570] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.958752] env[61991]: INFO nova.compute.manager [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1025.959236] env[61991]: DEBUG oslo.service.loopingcall [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.959431] env[61991]: DEBUG nova.compute.provider_tree [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.960585] env[61991]: DEBUG nova.compute.manager [-] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1025.960674] env[61991]: DEBUG nova.network.neutron [-] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1026.136045] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1026.150571] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130322, 'name': CreateVM_Task, 'duration_secs': 0.472059} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.150922] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.152246] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'guest_format': None, 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246962', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'name': 'volume-f529fecd-03fc-4258-a8dc-f770503a03ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c43df224-7c35-4b82-ba13-50f3a6f93f2e', 'attached_at': '', 'detached_at': '', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'serial': 'f529fecd-03fc-4258-a8dc-f770503a03ee'}, 'attachment_id': '5625e464-19bb-4cfc-9019-33372acf3876', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61991) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1026.153769] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Root volume attach. Driver type: vmdk {{(pid=61991) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1026.154200] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141efb28-b09b-4654-9733-651e262d0130 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.167029] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab3a48a-1cec-4069-bd53-ade3d7ea5601 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.173450] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a366b5-617e-43eb-89da-1651259fd5ff {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.181847] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-b93829d9-4b01-45bf-be65-31987c2c5c03 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.193025] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1026.193025] env[61991]: value = "task-1130325" [ 1026.193025] env[61991]: _type = "Task" [ 1026.193025] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.204414] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.297646] env[61991]: DEBUG oslo_vmware.api [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229217} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.298162] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.298604] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.299047] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.300995] env[61991]: INFO nova.compute.manager [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1026.300995] env[61991]: DEBUG oslo.service.loopingcall [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.300995] env[61991]: DEBUG nova.compute.manager [-] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.300995] env[61991]: DEBUG nova.network.neutron [-] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1026.464414] env[61991]: DEBUG nova.scheduler.client.report [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.468902] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updating instance_info_cache with network_info: [{"id": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "address": "fa:16:3e:8e:7c:67", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fdca45f-c9", "ovs_interfaceid": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.642357] env[61991]: DEBUG nova.compute.manager [req-0ffdb0f5-efac-42d2-8ecc-f3e399d79e72 req-ece428dc-cdd3-496f-bbab-6273cb99bdd3 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Received event network-vif-deleted-4e71b591-340e-4029-8776-b78a3dc981ce {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.643059] env[61991]: INFO nova.compute.manager [req-0ffdb0f5-efac-42d2-8ecc-f3e399d79e72 req-ece428dc-cdd3-496f-bbab-6273cb99bdd3 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Neutron deleted interface 4e71b591-340e-4029-8776-b78a3dc981ce; detaching it from the instance and deleting it from the info cache [ 1026.643059] env[61991]: DEBUG nova.network.neutron [req-0ffdb0f5-efac-42d2-8ecc-f3e399d79e72 req-ece428dc-cdd3-496f-bbab-6273cb99bdd3 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.672870] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.708194] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 34%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.874018] env[61991]: INFO nova.compute.manager [None req-ce53c168-ce34-45ed-b28a-df91d246e149 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance to original state: 'active' [ 1026.974028] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.974028] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.977187] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.977676] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1026.979113] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.895s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.979113] env[61991]: DEBUG nova.objects.instance [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'resources' on Instance uuid 5364f93c-ca53-480b-8397-c331dfdeed3f {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.981374] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.981637] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.981747] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.981895] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.982098] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.982196] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1026.982602] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1026.982602] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1027.108067] env[61991]: DEBUG nova.network.neutron [-] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.147436] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-778cdb3b-e15b-4080-90fd-719ca024769d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.165142] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4c14aa-1d54-4ebc-b202-23470bdbdb90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.208479] env[61991]: DEBUG nova.compute.manager [req-0ffdb0f5-efac-42d2-8ecc-f3e399d79e72 req-ece428dc-cdd3-496f-bbab-6273cb99bdd3 service nova] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Detach interface failed, port_id=4e71b591-340e-4029-8776-b78a3dc981ce, reason: Instance 892323a3-8661-47da-a52a-d7e0557ea300 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1027.211692] env[61991]: DEBUG nova.network.neutron [-] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.220302] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 47%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.479906] env[61991]: DEBUG nova.compute.utils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.481475] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1027.481875] env[61991]: DEBUG nova.network.neutron [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1027.495023] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.541819] env[61991]: DEBUG nova.policy [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceec39441db345a1ac9cc3f9d6f6ee18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '870d2c2c0e554180b190b88bdab5fc2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.610975] env[61991]: INFO nova.compute.manager [-] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Took 1.65 seconds to deallocate network for instance. [ 1027.707683] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 60%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.714198] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173bdf8f-bf0b-4566-a780-facb2b2a9fe1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.721758] env[61991]: INFO nova.compute.manager [-] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Took 1.42 seconds to deallocate network for instance. [ 1027.724345] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81d222a-529e-484f-a756-957cd62afa2c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.760174] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c357e8-1002-49c0-bee9-f17e13e84069 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.769509] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd5bb24-be76-4e55-b221-7fdc2a1b721f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.786095] env[61991]: DEBUG nova.compute.provider_tree [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.984831] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1028.035462] env[61991]: DEBUG nova.network.neutron [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Successfully created port: e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.123377] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.209060] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 73%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.234029] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.298025] env[61991]: DEBUG nova.scheduler.client.report [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.299974] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.300322] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.300570] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.300767] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.300945] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.304054] env[61991]: INFO nova.compute.manager [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Terminating instance [ 1028.307279] env[61991]: DEBUG nova.compute.manager [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.307597] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.307818] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d479c3a-9e4d-45fc-8117-6c4a63f0a43e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.317745] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1028.317745] env[61991]: value = "task-1130326" [ 1028.317745] env[61991]: _type = "Task" [ 1028.317745] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.333462] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.711525] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 86%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.753130] env[61991]: DEBUG nova.compute.manager [req-352e2b47-323e-419c-95bb-9a22b81731d5 req-1273f714-99ac-4bc0-a919-3fb4928e5f27 service nova] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Received event network-vif-deleted-15cc98a6-0818-4630-b737-e50e1fb71851 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.805379] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.808591] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.135s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.810241] env[61991]: INFO nova.compute.claims [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.821482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.821735] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.827841] env[61991]: INFO nova.scheduler.client.report [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleted allocations for instance 5364f93c-ca53-480b-8397-c331dfdeed3f [ 1028.836321] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.995092] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1029.024985] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.025310] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.025475] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.025733] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.025935] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.026471] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.026471] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.026724] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.026724] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.026957] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.027180] env[61991]: DEBUG nova.virt.hardware [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.028640] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb785fdc-199f-49a2-94b1-32749303046a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.042577] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab102e42-fb1b-478a-9356-afc8d4305b59 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.212155] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 97%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.335789] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1029.338499] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130326, 'name': PowerOffVM_Task, 'duration_secs': 0.748082} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.342108] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.342314] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1029.342514] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246948', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'name': 'volume-1b159140-cae6-4f5b-9849-8a3da257f4bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '1730f9b3-69ad-476b-b3ab-3e1345f2a115', 'attached_at': '2024-10-04T09:25:50.000000', 'detached_at': '', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'serial': '1b159140-cae6-4f5b-9849-8a3da257f4bb'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1029.343040] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7c0a26d-017d-4473-ba74-128a6dceb5f2 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "5364f93c-ca53-480b-8397-c331dfdeed3f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.177s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.344636] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a9be0a-bd0a-4eec-a479-ecabef31c60a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.369670] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec82ba24-4022-4b6f-84f4-6bbef2743e7e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.378365] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03cdc2b6-acd7-4c6a-bcc5-aded49fad7e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.401652] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52aa16f-5d6b-45ce-bde3-0b61fa3e45fe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.420293] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] The volume has not been displaced from its original location: [datastore1] volume-1b159140-cae6-4f5b-9849-8a3da257f4bb/volume-1b159140-cae6-4f5b-9849-8a3da257f4bb.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1029.425349] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1029.425683] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-874f8237-8ff4-4e60-ace4-a4977016584e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.443936] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1029.443936] env[61991]: value = "task-1130327" [ 1029.443936] env[61991]: _type = "Task" [ 1029.443936] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.452218] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130327, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.688023] env[61991]: DEBUG nova.network.neutron [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Successfully updated port: e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.712551] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task} progress is 98%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.862018] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.962390] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130327, 'name': ReconfigVM_Task, 'duration_secs': 0.242859} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.962390] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1029.969183] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dac268b2-13f5-440b-9008-807e47e64570 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.985856] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1029.985856] env[61991]: value = "task-1130328" [ 1029.985856] env[61991]: _type = "Task" [ 1029.985856] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.997814] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130328, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.031118] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07242b21-0bdc-4ea7-a7a0-95494c60711a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.040978] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23270309-54fa-43fc-87ac-396ba9fb1104 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.074702] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cc5e57-ad2e-45f0-bf56-03606ab4b934 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.083193] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f74b4d1-3af1-4592-8824-d907f810fdd7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.098931] env[61991]: DEBUG nova.compute.provider_tree [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.191051] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.191184] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.191277] env[61991]: DEBUG nova.network.neutron [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.214886] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130325, 'name': RelocateVM_Task, 'duration_secs': 3.94378} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.215194] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1030.215395] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246962', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'name': 'volume-f529fecd-03fc-4258-a8dc-f770503a03ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c43df224-7c35-4b82-ba13-50f3a6f93f2e', 'attached_at': '', 'detached_at': '', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'serial': 'f529fecd-03fc-4258-a8dc-f770503a03ee'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1030.216151] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d4f8fd-324f-496a-a88d-3d43c264f5e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.232750] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ea1189-96d6-4498-8639-3cffcc186815 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.256694] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-f529fecd-03fc-4258-a8dc-f770503a03ee/volume-f529fecd-03fc-4258-a8dc-f770503a03ee.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.257015] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ecec072-abc1-48f1-a493-cdda0a99dcc4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.276449] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1030.276449] env[61991]: value = "task-1130330" [ 1030.276449] env[61991]: _type = "Task" [ 1030.276449] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.284202] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130330, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.497605] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130328, 'name': ReconfigVM_Task, 'duration_secs': 0.19045} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.497932] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246948', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'name': 'volume-1b159140-cae6-4f5b-9849-8a3da257f4bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '1730f9b3-69ad-476b-b3ab-3e1345f2a115', 'attached_at': '2024-10-04T09:25:50.000000', 'detached_at': '', 'volume_id': '1b159140-cae6-4f5b-9849-8a3da257f4bb', 'serial': '1b159140-cae6-4f5b-9849-8a3da257f4bb'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1030.498233] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1030.498995] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd52e3a6-abfb-4bc7-bedb-3fdaf342023d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.505856] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.506102] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2de23380-38c4-4df2-b369-28fcb1a0575a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.585541] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.585676] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.585867] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleting the datastore file [datastore1] 1730f9b3-69ad-476b-b3ab-3e1345f2a115 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.586180] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a49069d-688c-4fe4-a89d-f4a110698a5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.593695] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1030.593695] env[61991]: value = "task-1130332" [ 1030.593695] env[61991]: _type = "Task" [ 1030.593695] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.603653] env[61991]: DEBUG nova.scheduler.client.report [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.606239] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.724335] env[61991]: DEBUG nova.network.neutron [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.779315] env[61991]: DEBUG nova.compute.manager [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Received event network-vif-plugged-e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.779639] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] Acquiring lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.779958] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.779958] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.780243] env[61991]: DEBUG nova.compute.manager [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] No waiting events found dispatching network-vif-plugged-e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.780323] env[61991]: WARNING nova.compute.manager [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Received unexpected event network-vif-plugged-e0d6f075-39ac-4264-9656-b78feb2b3747 for instance with vm_state building and task_state spawning. [ 1030.780486] env[61991]: DEBUG nova.compute.manager [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Received event network-changed-e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.780765] env[61991]: DEBUG nova.compute.manager [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Refreshing instance network info cache due to event network-changed-e0d6f075-39ac-4264-9656-b78feb2b3747. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.780840] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.791302] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130330, 'name': ReconfigVM_Task, 'duration_secs': 0.254739} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.791862] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-f529fecd-03fc-4258-a8dc-f770503a03ee/volume-f529fecd-03fc-4258-a8dc-f770503a03ee.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.798948] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25b4f27d-c364-4821-842a-592e10a7319e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.821630] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1030.821630] env[61991]: value = "task-1130333" [ 1030.821630] env[61991]: _type = "Task" [ 1030.821630] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.833988] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130333, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.890265] env[61991]: DEBUG nova.network.neutron [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.105934] env[61991]: DEBUG oslo_vmware.api [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147218} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.106233] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.106420] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.106646] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.106852] env[61991]: INFO nova.compute.manager [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Took 2.80 seconds to destroy the instance on the hypervisor. [ 1031.107111] env[61991]: DEBUG oslo.service.loopingcall [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.107926] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.108332] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.111020] env[61991]: DEBUG nova.compute.manager [-] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1031.111020] env[61991]: DEBUG nova.network.neutron [-] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1031.113168] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.621s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.113409] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.113498] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1031.113775] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.991s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.113981] env[61991]: DEBUG nova.objects.instance [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lazy-loading 'resources' on Instance uuid 892323a3-8661-47da-a52a-d7e0557ea300 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.116743] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467fd6f4-2486-4add-85bd-1eae9aa12684 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.124887] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf08cdd0-8889-4285-bcc6-80381f067929 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.140334] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbe5125-be86-4dc7-b0f6-6c0bee3c2a2f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.149782] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6757a459-9987-4095-929b-7a8d0349249c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.179288] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180254MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1031.179487] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.332925] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130333, 'name': ReconfigVM_Task, 'duration_secs': 0.155561} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.333352] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246962', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'name': 'volume-f529fecd-03fc-4258-a8dc-f770503a03ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c43df224-7c35-4b82-ba13-50f3a6f93f2e', 'attached_at': '', 'detached_at': '', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'serial': 'f529fecd-03fc-4258-a8dc-f770503a03ee'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1031.333908] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b3c5dc6-8eb4-4966-991c-3c40486037a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.341158] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1031.341158] env[61991]: value = "task-1130334" [ 1031.341158] env[61991]: _type = "Task" [ 1031.341158] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.349295] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130334, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.393112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.393485] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Instance network_info: |[{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.393801] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.393991] env[61991]: DEBUG nova.network.neutron [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Refreshing network info cache for port e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.395314] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:54:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0d6f075-39ac-4264-9656-b78feb2b3747', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.405879] env[61991]: DEBUG oslo.service.loopingcall [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.408317] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1031.408815] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ff2c727-e13f-4eba-814c-3f3c3576aa70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.434351] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.434351] env[61991]: value = "task-1130335" [ 1031.434351] env[61991]: _type = "Task" [ 1031.434351] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.445705] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130335, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.614990] env[61991]: DEBUG nova.compute.utils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.616560] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.616791] env[61991]: DEBUG nova.network.neutron [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1031.665386] env[61991]: DEBUG nova.policy [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6422dfa4c424fbaab78c3e72a0e6669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dddb51029854105bc6c9c8724181d39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1031.716943] env[61991]: DEBUG nova.network.neutron [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updated VIF entry in instance network info cache for port e0d6f075-39ac-4264-9656-b78feb2b3747. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1031.717367] env[61991]: DEBUG nova.network.neutron [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.802500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9614e1ef-95f2-4dad-ae01-fc4efe31e453 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.811118] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bbfc3f-90e4-4891-b312-f3164d6e661b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.842619] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c916a510-1650-4805-8592-c1d2c540c17e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.853964] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130334, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.857021] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd60dd0-b88c-4a79-adba-b55b789c3fa9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.870835] env[61991]: DEBUG nova.compute.provider_tree [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.912455] env[61991]: DEBUG nova.network.neutron [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Successfully created port: 3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.952285] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130335, 'name': CreateVM_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.121992] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.129913] env[61991]: DEBUG nova.network.neutron [-] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.220450] env[61991]: DEBUG oslo_concurrency.lockutils [req-ce1c206d-8470-4441-a7ae-1c24a5e0d7f1 req-eb012b4e-ca77-4b35-8ca7-26a579a46e0a service nova] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.357260] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130334, 'name': Rename_Task, 'duration_secs': 0.718661} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.357954] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.357954] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a538b24e-04a7-467e-a3d4-5e794cc27676 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.365779] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1032.365779] env[61991]: value = "task-1130336" [ 1032.365779] env[61991]: _type = "Task" [ 1032.365779] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.375016] env[61991]: DEBUG nova.scheduler.client.report [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.378308] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.445337] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130335, 'name': CreateVM_Task, 'duration_secs': 0.779987} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.445520] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1032.446224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.446413] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.446751] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1032.447098] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-544c5a4d-e953-4c7d-8f56-430fdca44911 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.451992] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1032.451992] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527a2ee9-4787-5645-2ba7-147d879963e4" [ 1032.451992] env[61991]: _type = "Task" [ 1032.451992] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.460215] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527a2ee9-4787-5645-2ba7-147d879963e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.633484] env[61991]: INFO nova.compute.manager [-] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Took 1.52 seconds to deallocate network for instance. [ 1032.807552] env[61991]: DEBUG nova.compute.manager [req-6a8138f7-87d9-4601-8a8f-aeab4e2c4f44 req-95511e2f-66b3-4c1b-b636-04e54040d16d service nova] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Received event network-vif-deleted-4e506178-06a6-49c6-a6b3-fa24ebafc988 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.876859] env[61991]: DEBUG oslo_vmware.api [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130336, 'name': PowerOnVM_Task, 'duration_secs': 0.445916} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.877156] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.877361] env[61991]: INFO nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Took 9.32 seconds to spawn the instance on the hypervisor. [ 1032.877544] env[61991]: DEBUG nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1032.878371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f803fa8-af42-40f4-91b3-1d9f0268ab34 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.881234] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.883361] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.650s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.883516] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.885168] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.026s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.886892] env[61991]: INFO nova.compute.claims [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1032.913828] env[61991]: INFO nova.scheduler.client.report [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Deleted allocations for instance 892323a3-8661-47da-a52a-d7e0557ea300 [ 1032.914989] env[61991]: INFO nova.scheduler.client.report [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocations for instance 8005b8f6-6f4d-4f72-be5d-955fa114408d [ 1032.963894] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]527a2ee9-4787-5645-2ba7-147d879963e4, 'name': SearchDatastore_Task, 'duration_secs': 0.009292} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.963894] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.963894] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.964567] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.964767] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.964968] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.966298] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99e3d740-1738-4509-a5ab-85653139205a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.974820] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.975016] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.975936] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-145bf6d1-d9da-4edb-a593-a7597b0eb344 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.981051] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1032.981051] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b69980-c444-24e3-2e66-4febd17cebfa" [ 1032.981051] env[61991]: _type = "Task" [ 1032.981051] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.988674] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b69980-c444-24e3-2e66-4febd17cebfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.132590] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.159891] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.160233] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.160401] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.160600] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.160753] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.160918] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.161154] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.161321] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.161503] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.161669] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.161858] env[61991]: DEBUG nova.virt.hardware [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.163410] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bb90cb-e2ad-4b99-a10e-16ffd59c4a60 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.171746] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fe22c9-d058-401a-b6c0-3746b8e1d158 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.177025] env[61991]: INFO nova.compute.manager [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Took 0.54 seconds to detach 1 volumes for instance. [ 1033.404150] env[61991]: INFO nova.compute.manager [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Took 18.09 seconds to build instance. [ 1033.426774] env[61991]: DEBUG oslo_concurrency.lockutils [None req-afaf126d-798b-44d9-bc67-b19e3d7d8361 tempest-ServersV294TestFqdnHostnames-1022215290 tempest-ServersV294TestFqdnHostnames-1022215290-project-member] Lock "892323a3-8661-47da-a52a-d7e0557ea300" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.668s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.428515] env[61991]: DEBUG oslo_concurrency.lockutils [None req-8dd4b841-04b3-43de-9934-cd1450967eb6 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "8005b8f6-6f4d-4f72-be5d-955fa114408d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.271s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.491174] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b69980-c444-24e3-2e66-4febd17cebfa, 'name': SearchDatastore_Task, 'duration_secs': 0.009937} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.491954] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0fe6b27-1abc-4c45-9570-4aebf6d108ea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.497193] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1033.497193] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520dd701-d77e-3d95-6823-6f9239858629" [ 1033.497193] env[61991]: _type = "Task" [ 1033.497193] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.505623] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520dd701-d77e-3d95-6823-6f9239858629, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.692219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.708587] env[61991]: DEBUG nova.compute.manager [req-89448bfe-77ff-48e9-8880-1a06fee12f48 req-c1581c69-dcc8-414a-8f5a-302c11e833b9 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Received event network-vif-plugged-3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.708707] env[61991]: DEBUG oslo_concurrency.lockutils [req-89448bfe-77ff-48e9-8880-1a06fee12f48 req-c1581c69-dcc8-414a-8f5a-302c11e833b9 service nova] Acquiring lock "81736f13-fc9e-4246-9d86-5a5bd7e15832-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.708808] env[61991]: DEBUG oslo_concurrency.lockutils [req-89448bfe-77ff-48e9-8880-1a06fee12f48 req-c1581c69-dcc8-414a-8f5a-302c11e833b9 service nova] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.709030] env[61991]: DEBUG oslo_concurrency.lockutils [req-89448bfe-77ff-48e9-8880-1a06fee12f48 req-c1581c69-dcc8-414a-8f5a-302c11e833b9 service nova] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.709163] env[61991]: DEBUG nova.compute.manager [req-89448bfe-77ff-48e9-8880-1a06fee12f48 req-c1581c69-dcc8-414a-8f5a-302c11e833b9 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] No waiting events found dispatching network-vif-plugged-3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.709499] env[61991]: WARNING nova.compute.manager [req-89448bfe-77ff-48e9-8880-1a06fee12f48 req-c1581c69-dcc8-414a-8f5a-302c11e833b9 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Received unexpected event network-vif-plugged-3a2b9da3-3552-419b-a28e-ed7f76eb9607 for instance with vm_state building and task_state spawning. [ 1033.904987] env[61991]: DEBUG nova.network.neutron [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Successfully updated port: 3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1033.906418] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ef996a7c-6a96-43af-955f-7cf4bedbb25d tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.605s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.008340] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520dd701-d77e-3d95-6823-6f9239858629, 'name': SearchDatastore_Task, 'duration_secs': 0.011033} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.011036] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.011331] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.012754] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db069723-db5b-4b23-a8c8-c8a1b9345c4e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.019412] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1034.019412] env[61991]: value = "task-1130337" [ 1034.019412] env[61991]: _type = "Task" [ 1034.019412] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.033253] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.073802] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8307f9bf-4864-4d74-9681-4d22b59c5f5c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.081585] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df87f95-979a-4c2f-b065-9b84650941b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.117570] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fa9c60-d549-4231-bce0-2351f1280981 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.125030] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fff1105-9889-4db2-9030-325d4db3613c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.138607] env[61991]: DEBUG nova.compute.provider_tree [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.408206] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "refresh_cache-81736f13-fc9e-4246-9d86-5a5bd7e15832" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.408362] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "refresh_cache-81736f13-fc9e-4246-9d86-5a5bd7e15832" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.408517] env[61991]: DEBUG nova.network.neutron [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.530084] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130337, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.644026] env[61991]: DEBUG nova.scheduler.client.report [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.810683] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "f7eab1a9-2a8a-4664-8736-ca599e207c31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.811498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.811498] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "f7eab1a9-2a8a-4664-8736-ca599e207c31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.811650] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.811907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.814584] env[61991]: INFO nova.compute.manager [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Terminating instance [ 1034.817930] env[61991]: DEBUG nova.compute.manager [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.818324] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1034.819807] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9266a0-09af-4538-9bdf-fd0ff76358de {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.834946] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.835400] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1ece1b7-9393-4b77-aaca-83a0ae403e95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.844735] env[61991]: DEBUG oslo_vmware.api [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1034.844735] env[61991]: value = "task-1130338" [ 1034.844735] env[61991]: _type = "Task" [ 1034.844735] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.855878] env[61991]: DEBUG oslo_vmware.api [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130338, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.859484] env[61991]: DEBUG nova.compute.manager [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Received event network-changed-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.859776] env[61991]: DEBUG nova.compute.manager [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Refreshing instance network info cache due to event network-changed-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1034.860060] env[61991]: DEBUG oslo_concurrency.lockutils [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] Acquiring lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.860216] env[61991]: DEBUG oslo_concurrency.lockutils [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] Acquired lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.860402] env[61991]: DEBUG nova.network.neutron [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Refreshing network info cache for port 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.966197] env[61991]: DEBUG nova.network.neutron [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1035.032538] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130337, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.898593} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.032538] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.032538] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.032538] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b848306-1aa3-4391-b6fc-a84b163983cc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.037778] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1035.037778] env[61991]: value = "task-1130339" [ 1035.037778] env[61991]: _type = "Task" [ 1035.037778] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.046026] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130339, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.155172] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.155703] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1035.158737] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.979s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.206113] env[61991]: DEBUG nova.network.neutron [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Updating instance_info_cache with network_info: [{"id": "3a2b9da3-3552-419b-a28e-ed7f76eb9607", "address": "fa:16:3e:03:dd:6e", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a2b9da3-35", "ovs_interfaceid": "3a2b9da3-3552-419b-a28e-ed7f76eb9607", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.357416] env[61991]: DEBUG oslo_vmware.api [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130338, 'name': PowerOffVM_Task, 'duration_secs': 0.342046} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.358010] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.358400] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.358708] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68297b24-bcc1-496b-ae67-5466c0fa3a86 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.419729] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.419894] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.420088] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleting the datastore file [datastore2] f7eab1a9-2a8a-4664-8736-ca599e207c31 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.420359] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6864cc24-443d-421f-b406-08f2eda72033 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.428825] env[61991]: DEBUG oslo_vmware.api [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1035.428825] env[61991]: value = "task-1130341" [ 1035.428825] env[61991]: _type = "Task" [ 1035.428825] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.436483] env[61991]: DEBUG oslo_vmware.api [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130341, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.549773] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130339, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078455} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.549773] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.549906] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021ace1b-cd24-4464-9b89-f67f306402a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.572305] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.575341] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c5875e7-06ff-410a-a2ea-667556f5e0e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.596272] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1035.596272] env[61991]: value = "task-1130342" [ 1035.596272] env[61991]: _type = "Task" [ 1035.596272] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.604654] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130342, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.665227] env[61991]: DEBUG nova.compute.utils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.665227] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1035.665227] env[61991]: DEBUG nova.network.neutron [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1035.709224] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "refresh_cache-81736f13-fc9e-4246-9d86-5a5bd7e15832" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.709490] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Instance network_info: |[{"id": "3a2b9da3-3552-419b-a28e-ed7f76eb9607", "address": "fa:16:3e:03:dd:6e", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a2b9da3-35", "ovs_interfaceid": "3a2b9da3-3552-419b-a28e-ed7f76eb9607", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.709927] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:dd:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a2b9da3-3552-419b-a28e-ed7f76eb9607', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.720416] env[61991]: DEBUG oslo.service.loopingcall [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.721406] env[61991]: DEBUG nova.network.neutron [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updated VIF entry in instance network info cache for port 8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1035.721785] env[61991]: DEBUG nova.network.neutron [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updating instance_info_cache with network_info: [{"id": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "address": "fa:16:3e:8e:7c:67", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fdca45f-c9", "ovs_interfaceid": "8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.723225] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1035.725653] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6b4fccb-a213-4168-a5e6-49f89e4da591 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.743828] env[61991]: DEBUG oslo_concurrency.lockutils [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] Releasing lock "refresh_cache-cf99c9cc-24c3-4acc-8120-49c4b12a3553" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.744118] env[61991]: DEBUG nova.compute.manager [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Received event network-changed-7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.744300] env[61991]: DEBUG nova.compute.manager [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Refreshing instance network info cache due to event network-changed-7277dc2e-f381-41e1-95de-31eda0f6cb27. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1035.744545] env[61991]: DEBUG oslo_concurrency.lockutils [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] Acquiring lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.744663] env[61991]: DEBUG oslo_concurrency.lockutils [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] Acquired lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.744826] env[61991]: DEBUG nova.network.neutron [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Refreshing network info cache for port 7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.751773] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.751773] env[61991]: value = "task-1130343" [ 1035.751773] env[61991]: _type = "Task" [ 1035.751773] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.758063] env[61991]: DEBUG nova.compute.manager [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Received event network-changed-3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.758281] env[61991]: DEBUG nova.compute.manager [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Refreshing instance network info cache due to event network-changed-3a2b9da3-3552-419b-a28e-ed7f76eb9607. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1035.758494] env[61991]: DEBUG oslo_concurrency.lockutils [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] Acquiring lock "refresh_cache-81736f13-fc9e-4246-9d86-5a5bd7e15832" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.758648] env[61991]: DEBUG oslo_concurrency.lockutils [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] Acquired lock "refresh_cache-81736f13-fc9e-4246-9d86-5a5bd7e15832" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.758805] env[61991]: DEBUG nova.network.neutron [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Refreshing network info cache for port 3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.765793] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130343, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.776943] env[61991]: DEBUG nova.policy [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1035.843093] env[61991]: DEBUG nova.compute.manager [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1035.938846] env[61991]: DEBUG oslo_vmware.api [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130341, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.353199} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.939565] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.939565] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1035.939748] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1035.939811] env[61991]: INFO nova.compute.manager [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1035.940078] env[61991]: DEBUG oslo.service.loopingcall [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.940411] env[61991]: DEBUG nova.compute.manager [-] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.940411] env[61991]: DEBUG nova.network.neutron [-] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1036.108017] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130342, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.175014] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1036.181350] env[61991]: INFO nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating resource usage from migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd [ 1036.190249] env[61991]: DEBUG nova.network.neutron [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Successfully created port: 7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.207029] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.207344] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.207344] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f7eab1a9-2a8a-4664-8736-ca599e207c31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.207485] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 100d5fc0-e9d9-4892-9f60-99d657c533ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 5be79140-21f9-493c-a840-2d905b8d66bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.209185] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 1730f9b3-69ad-476b-b3ab-3e1345f2a115 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance b26f02c6-a8ea-402d-922d-7879ccb28b92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 81736f13-fc9e-4246-9d86-5a5bd7e15832 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 84a0dee7-c33b-494a-ad38-83da6ab44ce2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c43df224-7c35-4b82-ba13-50f3a6f93f2e actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1036.209185] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1036.271811] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130343, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.359496] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.387552] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4ef8fa-a8f9-4915-a9af-a6a1ab9ef1fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.398626] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b152ab-9737-4919-9d40-4af9cde2330d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.439846] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029cd75f-a0db-41e8-ba5f-d11921a9db46 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.447921] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aba9684-cca0-4e2f-a852-de7950c726ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.463035] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.538064] env[61991]: DEBUG nova.network.neutron [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Updated VIF entry in instance network info cache for port 3a2b9da3-3552-419b-a28e-ed7f76eb9607. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.538450] env[61991]: DEBUG nova.network.neutron [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Updating instance_info_cache with network_info: [{"id": "3a2b9da3-3552-419b-a28e-ed7f76eb9607", "address": "fa:16:3e:03:dd:6e", "network": {"id": "b75095f9-5f22-4806-bbf5-f0a13d04b146", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-800261984-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dddb51029854105bc6c9c8724181d39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a2b9da3-35", "ovs_interfaceid": "3a2b9da3-3552-419b-a28e-ed7f76eb9607", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.616364] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130342, 'name': ReconfigVM_Task, 'duration_secs': 0.672595} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.617797] env[61991]: DEBUG nova.network.neutron [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updated VIF entry in instance network info cache for port 7277dc2e-f381-41e1-95de-31eda0f6cb27. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.618343] env[61991]: DEBUG nova.network.neutron [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [{"id": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "address": "fa:16:3e:37:4b:7d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7277dc2e-f3", "ovs_interfaceid": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.620062] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.621172] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23c93dbc-ab16-48cd-be46-693a2eac1531 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.629154] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1036.629154] env[61991]: value = "task-1130344" [ 1036.629154] env[61991]: _type = "Task" [ 1036.629154] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.638781] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130344, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.764552] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130343, 'name': CreateVM_Task, 'duration_secs': 0.519641} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.764669] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1036.765368] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.765541] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.765875] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.766156] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836d1a6a-a54c-4dcf-bf54-be8da1ba30dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.770816] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1036.770816] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b30e2-4ac3-dbfa-076b-abf8f0b522d8" [ 1036.770816] env[61991]: _type = "Task" [ 1036.770816] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.778913] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b30e2-4ac3-dbfa-076b-abf8f0b522d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.810313] env[61991]: DEBUG nova.network.neutron [-] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.965677] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.041573] env[61991]: DEBUG oslo_concurrency.lockutils [req-51f7683e-2948-4ba7-8511-386f7e2d6e4e req-36c63e92-c05f-4890-b899-8f8a1671d066 service nova] Releasing lock "refresh_cache-81736f13-fc9e-4246-9d86-5a5bd7e15832" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.122264] env[61991]: DEBUG oslo_concurrency.lockutils [req-fff8d2c9-c97c-40a4-b476-0c1c451c813c req-267fe637-a45f-4ce4-bb50-db9c594dab5b service nova] Releasing lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.140046] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130344, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.189485] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1037.215163] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.215442] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.215607] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.215795] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.215947] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.216112] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.216318] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.216480] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.216707] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.217739] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.217977] env[61991]: DEBUG nova.virt.hardware [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.218975] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7784546-5384-4dc7-987e-907e90580ffe {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.226995] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da645b5-40ba-478a-bdd9-a3e936affcaf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.281248] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]529b30e2-4ac3-dbfa-076b-abf8f0b522d8, 'name': SearchDatastore_Task, 'duration_secs': 0.01464} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.281594] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.281892] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.282191] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.282594] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.282594] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.282850] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc51cd96-3f43-4d8b-90a4-da55c7c0e503 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.294933] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.295126] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1037.295824] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9de7c616-02c0-40dc-8e12-9e4d09a6e4a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.301489] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1037.301489] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d8a4bd-c8ad-7c83-2e02-f99120f1acb2" [ 1037.301489] env[61991]: _type = "Task" [ 1037.301489] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.309161] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d8a4bd-c8ad-7c83-2e02-f99120f1acb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.314668] env[61991]: INFO nova.compute.manager [-] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Took 1.37 seconds to deallocate network for instance. [ 1037.471320] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1037.471542] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.313s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.471821] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.780s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.472017] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.474084] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.115s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.498740] env[61991]: INFO nova.scheduler.client.report [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted allocations for instance 1730f9b3-69ad-476b-b3ab-3e1345f2a115 [ 1037.640058] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130344, 'name': Rename_Task, 'duration_secs': 0.755867} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.640365] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.640615] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f60ba7d-ab64-4576-8248-59c9bd578995 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.647570] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1037.647570] env[61991]: value = "task-1130345" [ 1037.647570] env[61991]: _type = "Task" [ 1037.647570] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.655427] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.746271] env[61991]: DEBUG nova.compute.manager [req-f68e2ad1-56fc-4ca8-895a-feaaab0c09d2 req-04eb40e3-ea2d-46dc-9651-f363279bacc4 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Received event network-vif-plugged-7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.746271] env[61991]: DEBUG oslo_concurrency.lockutils [req-f68e2ad1-56fc-4ca8-895a-feaaab0c09d2 req-04eb40e3-ea2d-46dc-9651-f363279bacc4 service nova] Acquiring lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.746425] env[61991]: DEBUG oslo_concurrency.lockutils [req-f68e2ad1-56fc-4ca8-895a-feaaab0c09d2 req-04eb40e3-ea2d-46dc-9651-f363279bacc4 service nova] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.746561] env[61991]: DEBUG oslo_concurrency.lockutils [req-f68e2ad1-56fc-4ca8-895a-feaaab0c09d2 req-04eb40e3-ea2d-46dc-9651-f363279bacc4 service nova] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.746810] env[61991]: DEBUG nova.compute.manager [req-f68e2ad1-56fc-4ca8-895a-feaaab0c09d2 req-04eb40e3-ea2d-46dc-9651-f363279bacc4 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] No waiting events found dispatching network-vif-plugged-7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1037.747098] env[61991]: WARNING nova.compute.manager [req-f68e2ad1-56fc-4ca8-895a-feaaab0c09d2 req-04eb40e3-ea2d-46dc-9651-f363279bacc4 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Received unexpected event network-vif-plugged-7f4f6f32-7ee7-4884-86c4-27ecf7daca85 for instance with vm_state building and task_state spawning. [ 1037.785371] env[61991]: DEBUG nova.compute.manager [req-705e4d85-fdbb-48be-94ea-12963777dab9 req-f706b247-29cf-49ab-8ee3-7e82745739a7 service nova] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Received event network-vif-deleted-ce8085ee-47de-4c70-aafc-dcdf66c17383 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.815604] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d8a4bd-c8ad-7c83-2e02-f99120f1acb2, 'name': SearchDatastore_Task, 'duration_secs': 0.011834} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.816424] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7917e67-b2fd-4cb9-850d-2f7af219c561 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.820300] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.823323] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1037.823323] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d99239-3388-9294-eb1d-0d5f090d0f54" [ 1037.823323] env[61991]: _type = "Task" [ 1037.823323] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.831295] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d99239-3388-9294-eb1d-0d5f090d0f54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.850887] env[61991]: DEBUG nova.network.neutron [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Successfully updated port: 7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.979634] env[61991]: INFO nova.compute.claims [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.007047] env[61991]: DEBUG oslo_concurrency.lockutils [None req-20179d00-0265-4294-a8e5-3eefae6f62fb tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "1730f9b3-69ad-476b-b3ab-3e1345f2a115" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.706s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.158704] env[61991]: DEBUG oslo_vmware.api [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130345, 'name': PowerOnVM_Task, 'duration_secs': 0.448025} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.158985] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.159219] env[61991]: INFO nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Took 9.16 seconds to spawn the instance on the hypervisor. [ 1038.159436] env[61991]: DEBUG nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.160221] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7857592-161e-418a-8004-73f70b21a9f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.334803] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d99239-3388-9294-eb1d-0d5f090d0f54, 'name': SearchDatastore_Task, 'duration_secs': 0.0085} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.335112] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.335375] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 81736f13-fc9e-4246-9d86-5a5bd7e15832/81736f13-fc9e-4246-9d86-5a5bd7e15832.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.335661] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7abda83-fe4b-4faa-9333-3e4aa4e6792b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.346027] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1038.346027] env[61991]: value = "task-1130346" [ 1038.346027] env[61991]: _type = "Task" [ 1038.346027] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.352822] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130346, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.352983] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-84a0dee7-c33b-494a-ad38-83da6ab44ce2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.353161] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-84a0dee7-c33b-494a-ad38-83da6ab44ce2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.353268] env[61991]: DEBUG nova.network.neutron [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1038.486071] env[61991]: INFO nova.compute.resource_tracker [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating resource usage from migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd [ 1038.632892] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb687a0-34d4-477b-ab2c-a7cecbc99edc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.643235] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca04a6f-1d43-4e3e-a249-e78fb137d6c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.687081] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee97b8db-169b-49b1-b3dd-3be21d445914 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.693250] env[61991]: INFO nova.compute.manager [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Took 14.02 seconds to build instance. [ 1038.700594] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daed34bb-ccda-4100-ac1a-de3f7d27a406 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.720973] env[61991]: DEBUG nova.compute.provider_tree [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.854407] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130346, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.888282] env[61991]: DEBUG nova.network.neutron [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1038.997628] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.997984] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.045789] env[61991]: DEBUG nova.network.neutron [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Updating instance_info_cache with network_info: [{"id": "7f4f6f32-7ee7-4884-86c4-27ecf7daca85", "address": "fa:16:3e:30:d7:6e", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f4f6f32-7e", "ovs_interfaceid": "7f4f6f32-7ee7-4884-86c4-27ecf7daca85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.196154] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b4509939-58b9-486d-8648-e9343f477b2d tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.550s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.224727] env[61991]: DEBUG nova.scheduler.client.report [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.355237] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130346, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548499} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.355577] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 81736f13-fc9e-4246-9d86-5a5bd7e15832/81736f13-fc9e-4246-9d86-5a5bd7e15832.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.355811] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.356157] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2858e316-f3e2-4317-b001-bb76e566d7e2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.364225] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1039.364225] env[61991]: value = "task-1130347" [ 1039.364225] env[61991]: _type = "Task" [ 1039.364225] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.372350] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.405949] env[61991]: DEBUG oslo_concurrency.lockutils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.406182] env[61991]: DEBUG oslo_concurrency.lockutils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.500482] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.547747] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-84a0dee7-c33b-494a-ad38-83da6ab44ce2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.547842] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Instance network_info: |[{"id": "7f4f6f32-7ee7-4884-86c4-27ecf7daca85", "address": "fa:16:3e:30:d7:6e", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f4f6f32-7e", "ovs_interfaceid": "7f4f6f32-7ee7-4884-86c4-27ecf7daca85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.548274] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:d7:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f4f6f32-7ee7-4884-86c4-27ecf7daca85', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.556245] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating folder: Project (07a063bdfecb4cf7956b078c86f239f2). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1039.556870] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5876774b-67ad-4fb2-9d0e-494d15ffd58b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.569263] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created folder: Project (07a063bdfecb4cf7956b078c86f239f2) in parent group-v246753. [ 1039.569466] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating folder: Instances. Parent ref: group-v246966. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1039.569735] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8012dbd-5810-4e86-ad26-4a519fecfb9c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.580494] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created folder: Instances in parent group-v246966. [ 1039.580767] env[61991]: DEBUG oslo.service.loopingcall [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.580976] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1039.581196] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c720b3fd-babc-4814-93a1-84ebd4dd86a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.601043] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.601043] env[61991]: value = "task-1130350" [ 1039.601043] env[61991]: _type = "Task" [ 1039.601043] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.609458] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130350, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.729892] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.256s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.730143] env[61991]: INFO nova.compute.manager [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Migrating [ 1039.736497] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.916s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.736785] env[61991]: DEBUG nova.objects.instance [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'resources' on Instance uuid f7eab1a9-2a8a-4664-8736-ca599e207c31 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.770738] env[61991]: DEBUG nova.compute.manager [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Received event network-changed-7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.771026] env[61991]: DEBUG nova.compute.manager [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Refreshing instance network info cache due to event network-changed-7f4f6f32-7ee7-4884-86c4-27ecf7daca85. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.771240] env[61991]: DEBUG oslo_concurrency.lockutils [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] Acquiring lock "refresh_cache-84a0dee7-c33b-494a-ad38-83da6ab44ce2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.771343] env[61991]: DEBUG oslo_concurrency.lockutils [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] Acquired lock "refresh_cache-84a0dee7-c33b-494a-ad38-83da6ab44ce2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.771513] env[61991]: DEBUG nova.network.neutron [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Refreshing network info cache for port 7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.878308] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130347, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068683} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.878452] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.879472] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e401ef0-6950-4e34-bf7e-3c877a7c6546 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.905688] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 81736f13-fc9e-4246-9d86-5a5bd7e15832/81736f13-fc9e-4246-9d86-5a5bd7e15832.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.906588] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f043644-ee23-4c33-9cea-f444afbf37ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.924580] env[61991]: DEBUG nova.compute.utils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.933897] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1039.933897] env[61991]: value = "task-1130351" [ 1039.933897] env[61991]: _type = "Task" [ 1039.933897] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.944832] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.980601] env[61991]: DEBUG nova.compute.manager [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Received event network-changed-e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.980801] env[61991]: DEBUG nova.compute.manager [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Refreshing instance network info cache due to event network-changed-e0d6f075-39ac-4264-9656-b78feb2b3747. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.981051] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.981237] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.981775] env[61991]: DEBUG nova.network.neutron [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Refreshing network info cache for port e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1040.022322] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.116438] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130350, 'name': CreateVM_Task, 'duration_secs': 0.382208} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.116913] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1040.117787] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.118161] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.118680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.119094] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f71ee8aa-511e-4467-b021-32dca1fedc21 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.126410] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1040.126410] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522b9b72-d384-2902-3107-71e9f14b010a" [ 1040.126410] env[61991]: _type = "Task" [ 1040.126410] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.136568] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522b9b72-d384-2902-3107-71e9f14b010a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.249563] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.249962] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.249962] env[61991]: DEBUG nova.network.neutron [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1040.400074] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5696b5-8d93-4161-864a-d9d4680c9038 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.408498] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f341450-50bd-45a8-81c2-5e8c8ec8737e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.441063] env[61991]: DEBUG oslo_concurrency.lockutils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.034s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.446475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5aeef34-ee69-4e52-8608-9c7b9ddc9876 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.460930] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.462792] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d0b1f0-38d8-4c30-8f9b-1baccbfc43bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.487176] env[61991]: DEBUG nova.compute.provider_tree [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.629225] env[61991]: DEBUG nova.network.neutron [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Updated VIF entry in instance network info cache for port 7f4f6f32-7ee7-4884-86c4-27ecf7daca85. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.629225] env[61991]: DEBUG nova.network.neutron [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Updating instance_info_cache with network_info: [{"id": "7f4f6f32-7ee7-4884-86c4-27ecf7daca85", "address": "fa:16:3e:30:d7:6e", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f4f6f32-7e", "ovs_interfaceid": "7f4f6f32-7ee7-4884-86c4-27ecf7daca85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.642314] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522b9b72-d384-2902-3107-71e9f14b010a, 'name': SearchDatastore_Task, 'duration_secs': 0.010718} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.643025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.643493] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.643943] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.644604] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.644604] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.644998] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce4c7bfd-cf7d-4b8d-bb51-3ad0f9114559 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.660067] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.660067] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1040.660781] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c50f13a-e135-4c5b-9bc2-2574f389d30d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.667222] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1040.667222] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d886cf-b294-b1d9-222a-6fc41ea4f879" [ 1040.667222] env[61991]: _type = "Task" [ 1040.667222] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.677055] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d886cf-b294-b1d9-222a-6fc41ea4f879, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.856208] env[61991]: DEBUG nova.network.neutron [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updated VIF entry in instance network info cache for port e0d6f075-39ac-4264-9656-b78feb2b3747. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.856707] env[61991]: DEBUG nova.network.neutron [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.953468] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130351, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.994503] env[61991]: DEBUG nova.scheduler.client.report [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.038694] env[61991]: DEBUG nova.network.neutron [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [{"id": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "address": "fa:16:3e:37:4b:7d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7277dc2e-f3", "ovs_interfaceid": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.136813] env[61991]: DEBUG oslo_concurrency.lockutils [req-aaf9d9d1-f4db-42f1-addd-c7908aa9d36d req-253cd106-6921-426b-8556-bb0c58254753 service nova] Releasing lock "refresh_cache-84a0dee7-c33b-494a-ad38-83da6ab44ce2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.179703] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d886cf-b294-b1d9-222a-6fc41ea4f879, 'name': SearchDatastore_Task, 'duration_secs': 0.011715} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.180493] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f70fc7a7-39a7-4402-ad21-8b0b849591fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.185541] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1041.185541] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5231621c-1b3c-1716-8217-4250921f2dff" [ 1041.185541] env[61991]: _type = "Task" [ 1041.185541] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.193211] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5231621c-1b3c-1716-8217-4250921f2dff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.360063] env[61991]: DEBUG oslo_concurrency.lockutils [req-e6c98859-30f1-4433-a4be-47ba80b4aaf9 req-19d58653-873b-4172-93cb-b4363454e86c service nova] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.452056] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130351, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.493903] env[61991]: DEBUG oslo_concurrency.lockutils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.494210] env[61991]: DEBUG oslo_concurrency.lockutils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.494451] env[61991]: INFO nova.compute.manager [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Attaching volume 69b6847b-fc53-4491-a7c8-08ee506cb1f9 to /dev/sdb [ 1041.498908] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.503026] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.481s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.504405] env[61991]: INFO nova.compute.claims [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.517996] env[61991]: INFO nova.scheduler.client.report [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleted allocations for instance f7eab1a9-2a8a-4664-8736-ca599e207c31 [ 1041.533894] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fa8d17-5ed5-46bb-988c-537ce7340c33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.542070] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.544054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d352ff2e-b65a-4a45-820b-0e83d2dc50dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.559917] env[61991]: DEBUG nova.virt.block_device [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updating existing volume attachment record: f06c0d34-7084-49b3-a8bc-acbf82cd282c {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1041.696685] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5231621c-1b3c-1716-8217-4250921f2dff, 'name': SearchDatastore_Task, 'duration_secs': 0.037011} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.696986] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.697343] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 84a0dee7-c33b-494a-ad38-83da6ab44ce2/84a0dee7-c33b-494a-ad38-83da6ab44ce2.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1041.697628] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5cd339b-b6d0-42af-8bc6-9a43bfd1ba08 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.706584] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1041.706584] env[61991]: value = "task-1130352" [ 1041.706584] env[61991]: _type = "Task" [ 1041.706584] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.714038] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130352, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.955396] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130351, 'name': ReconfigVM_Task, 'duration_secs': 1.646871} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.955725] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 81736f13-fc9e-4246-9d86-5a5bd7e15832/81736f13-fc9e-4246-9d86-5a5bd7e15832.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.956578] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f753de9-957f-4d1f-a985-74e05a90bf70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.964926] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1041.964926] env[61991]: value = "task-1130354" [ 1041.964926] env[61991]: _type = "Task" [ 1041.964926] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.974797] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130354, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.025970] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ec153d9e-2bbf-4d24-93e4-da3c0674b083 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "f7eab1a9-2a8a-4664-8736-ca599e207c31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.215s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.217917] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130352, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48258} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.218193] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 84a0dee7-c33b-494a-ad38-83da6ab44ce2/84a0dee7-c33b-494a-ad38-83da6ab44ce2.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.218416] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.218962] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e9cb2d4e-05c7-48ad-b076-87de951d1d6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.226854] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1042.226854] env[61991]: value = "task-1130355" [ 1042.226854] env[61991]: _type = "Task" [ 1042.226854] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.238353] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130355, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.475563] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130354, 'name': Rename_Task, 'duration_secs': 0.327577} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.475864] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.476130] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acd6733e-62b1-43db-a691-c4907206ba12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.483278] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1042.483278] env[61991]: value = "task-1130356" [ 1042.483278] env[61991]: _type = "Task" [ 1042.483278] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.493412] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.643482] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.643933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.644040] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.644368] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.644368] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.647995] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bee440-1e02-4521-b456-a2f60807ec14 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.651034] env[61991]: INFO nova.compute.manager [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Terminating instance [ 1042.653246] env[61991]: DEBUG nova.compute.manager [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1042.653518] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.654221] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a5a7cc-2187-47f0-acd1-21984e2236ef {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.659734] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f6fe89-66a0-4d21-a1b7-f4590ea8589e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.665205] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.665720] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e7589cd-8dda-4aee-b11c-8bcd54931899 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.694054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8eab638-42c2-4e60-9853-c16cc5af6783 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.696783] env[61991]: DEBUG oslo_vmware.api [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1042.696783] env[61991]: value = "task-1130357" [ 1042.696783] env[61991]: _type = "Task" [ 1042.696783] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.704179] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1582bf-7666-4699-bf6a-ba260653390f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.710737] env[61991]: DEBUG oslo_vmware.api [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130357, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.721746] env[61991]: DEBUG nova.compute.provider_tree [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.736283] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130355, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064429} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.736545] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.737337] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924bfcd6-2533-4ace-8030-318dad7bebb9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.759842] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 84a0dee7-c33b-494a-ad38-83da6ab44ce2/84a0dee7-c33b-494a-ad38-83da6ab44ce2.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.760144] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe8144e4-a8c7-479f-af6f-c7f0b491f2a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.781613] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1042.781613] env[61991]: value = "task-1130358" [ 1042.781613] env[61991]: _type = "Task" [ 1042.781613] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.789415] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130358, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.994703] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130356, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.065069] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e87e49f-b65b-4695-9ae8-f4e8b78155f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.086229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1043.207556] env[61991]: DEBUG oslo_vmware.api [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130357, 'name': PowerOffVM_Task, 'duration_secs': 0.260257} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.207877] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.208104] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.208396] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-849d3935-7cc3-49ca-aa28-8e4af0614a54 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.224447] env[61991]: DEBUG nova.scheduler.client.report [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.287567] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1043.287804] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1043.288066] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleting the datastore file [datastore2] 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.288799] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-619fec5d-dfd9-42f4-b10c-688989e33268 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.294400] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130358, 'name': ReconfigVM_Task, 'duration_secs': 0.294881} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.295078] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 84a0dee7-c33b-494a-ad38-83da6ab44ce2/84a0dee7-c33b-494a-ad38-83da6ab44ce2.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.295696] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3dc7be7-afd6-45eb-8ae3-87e7c8fa2171 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.303236] env[61991]: DEBUG oslo_vmware.api [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for the task: (returnval){ [ 1043.303236] env[61991]: value = "task-1130360" [ 1043.303236] env[61991]: _type = "Task" [ 1043.303236] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.303551] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1043.303551] env[61991]: value = "task-1130361" [ 1043.303551] env[61991]: _type = "Task" [ 1043.303551] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.314405] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130361, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.317270] env[61991]: DEBUG oslo_vmware.api [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130360, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.496353] env[61991]: DEBUG oslo_vmware.api [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130356, 'name': PowerOnVM_Task, 'duration_secs': 0.812619} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.496744] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1043.496856] env[61991]: INFO nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Took 10.36 seconds to spawn the instance on the hypervisor. [ 1043.497056] env[61991]: DEBUG nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1043.497781] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90eb09ec-29a0-4aa2-9193-b68589d73530 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.596313] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.596631] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29cb4738-0997-4b9d-85d3-63c6ebe0583a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.604905] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1043.604905] env[61991]: value = "task-1130362" [ 1043.604905] env[61991]: _type = "Task" [ 1043.604905] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.614307] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130362, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.729192] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.226s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.729741] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.817213] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130361, 'name': Rename_Task, 'duration_secs': 0.140185} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.820506] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.820741] env[61991]: DEBUG oslo_vmware.api [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Task: {'id': task-1130360, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.480878} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.821024] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93dc7b8d-afd0-43e5-b576-61202b4f3dda {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.822611] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.822816] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.822986] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.823185] env[61991]: INFO nova.compute.manager [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1043.823424] env[61991]: DEBUG oslo.service.loopingcall [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.823623] env[61991]: DEBUG nova.compute.manager [-] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1043.823716] env[61991]: DEBUG nova.network.neutron [-] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1043.831433] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1043.831433] env[61991]: value = "task-1130364" [ 1043.831433] env[61991]: _type = "Task" [ 1043.831433] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.844750] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130364, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.016165] env[61991]: INFO nova.compute.manager [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Took 17.38 seconds to build instance. [ 1044.086752] env[61991]: DEBUG nova.compute.manager [req-cc6fe248-704b-45be-8066-c738b412b03e req-94aec1b5-997a-4a02-b80e-94c3e7ee6ee0 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Received event network-vif-deleted-db6f94dd-1023-4cf6-8997-5bead8bfa1b2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.087603] env[61991]: INFO nova.compute.manager [req-cc6fe248-704b-45be-8066-c738b412b03e req-94aec1b5-997a-4a02-b80e-94c3e7ee6ee0 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Neutron deleted interface db6f94dd-1023-4cf6-8997-5bead8bfa1b2; detaching it from the instance and deleting it from the info cache [ 1044.087603] env[61991]: DEBUG nova.network.neutron [req-cc6fe248-704b-45be-8066-c738b412b03e req-94aec1b5-997a-4a02-b80e-94c3e7ee6ee0 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.115467] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130362, 'name': PowerOffVM_Task, 'duration_secs': 0.285963} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.115786] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1044.116023] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1044.235411] env[61991]: DEBUG nova.compute.utils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.236933] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1044.237116] env[61991]: DEBUG nova.network.neutron [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1044.281750] env[61991]: DEBUG nova.policy [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c5daf21b21b40c9a54bed64a64538bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '822d7e3c678e4defa24bb4d8439a62de', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.342271] env[61991]: DEBUG oslo_vmware.api [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130364, 'name': PowerOnVM_Task, 'duration_secs': 0.482556} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.342517] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.342732] env[61991]: INFO nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1044.342918] env[61991]: DEBUG nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.343756] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e8f195-0eff-4f79-a475-7d3e4f8b5a70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.519117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-524dd4d2-556a-403e-aa50-1b8f7ac91412 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.888s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.566206] env[61991]: DEBUG nova.network.neutron [-] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.589891] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e48f39f3-4f93-40a7-9a3a-232225e77153 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.603750] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0884d797-2299-4172-ae85-422ae6bf1e86 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.622611] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.622858] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.623021] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.623207] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.623352] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.623494] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.623694] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.623842] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.623998] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.624399] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.625039] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.639227] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2091f3a1-8aa8-4e7a-88e4-ae9bc8a53c51 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.651220] env[61991]: DEBUG nova.compute.manager [req-cc6fe248-704b-45be-8066-c738b412b03e req-94aec1b5-997a-4a02-b80e-94c3e7ee6ee0 service nova] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Detach interface failed, port_id=db6f94dd-1023-4cf6-8997-5bead8bfa1b2, reason: Instance 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1044.658690] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1044.658690] env[61991]: value = "task-1130365" [ 1044.658690] env[61991]: _type = "Task" [ 1044.658690] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.667977] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130365, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.708330] env[61991]: DEBUG nova.network.neutron [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Successfully created port: 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.740942] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.862788] env[61991]: INFO nova.compute.manager [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Took 15.02 seconds to build instance. [ 1045.070447] env[61991]: INFO nova.compute.manager [-] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Took 1.25 seconds to deallocate network for instance. [ 1045.168987] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130365, 'name': ReconfigVM_Task, 'duration_secs': 0.165469} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.169423] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1045.365468] env[61991]: DEBUG oslo_concurrency.lockutils [None req-a9490671-145a-4919-92e5-6aefb8f2dc00 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.543s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.559565] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "81736f13-fc9e-4246-9d86-5a5bd7e15832" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.559906] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.560155] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "81736f13-fc9e-4246-9d86-5a5bd7e15832-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.560470] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.560730] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.563263] env[61991]: INFO nova.compute.manager [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Terminating instance [ 1045.565259] env[61991]: DEBUG nova.compute.manager [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.565468] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.566341] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2587c5-4656-4d81-be31-25dbd5207ed8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.574570] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.574828] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48460197-9e14-4954-a154-52835ea3b964 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.577214] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.577448] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.577703] env[61991]: DEBUG nova.objects.instance [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lazy-loading 'resources' on Instance uuid 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.583777] env[61991]: DEBUG oslo_vmware.api [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1045.583777] env[61991]: value = "task-1130366" [ 1045.583777] env[61991]: _type = "Task" [ 1045.583777] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.595348] env[61991]: DEBUG oslo_vmware.api [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130366, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.676043] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.676043] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.676311] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.676368] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.677035] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.677035] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.677035] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.677211] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.677406] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.677617] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.677889] env[61991]: DEBUG nova.virt.hardware [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.683248] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1045.683607] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7df96c8c-bf0a-45f2-96fd-b37d4632ea39 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.702705] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1045.702705] env[61991]: value = "task-1130367" [ 1045.702705] env[61991]: _type = "Task" [ 1045.702705] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.711302] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.754343] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1045.774682] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.775020] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.775206] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.775414] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.775582] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.775749] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.775974] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.776159] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.776447] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.776575] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.776780] env[61991]: DEBUG nova.virt.hardware [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.777720] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37774833-4a82-4740-b8bb-dd0ba4809577 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.786157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bb552f-1371-4ee5-9242-ae8aa60fb4ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.092528] env[61991]: DEBUG oslo_vmware.api [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130366, 'name': PowerOffVM_Task, 'duration_secs': 0.255301} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.092800] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1046.093097] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1046.093247] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0436672e-810f-43f0-8a70-b14dd4a8b510 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.107343] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1046.107574] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246969', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'name': 'volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be79140-21f9-493c-a840-2d905b8d66bf', 'attached_at': '', 'detached_at': '', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'serial': '69b6847b-fc53-4491-a7c8-08ee506cb1f9'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1046.108762] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e490ec-b47e-4534-b012-7add38e2b2d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.129295] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a94ee2-fdbc-40cb-b7fb-e19a3ca04689 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.157475] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9/volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.159785] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa48212e-be60-488c-9e30-46b58985c0d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.174252] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.174252] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.174252] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleting the datastore file [datastore1] 81736f13-fc9e-4246-9d86-5a5bd7e15832 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.177651] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73ffbd2d-9fc3-4dd7-bc25-c15f3af394dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.179088] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1046.179088] env[61991]: value = "task-1130369" [ 1046.179088] env[61991]: _type = "Task" [ 1046.179088] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.185133] env[61991]: DEBUG oslo_vmware.api [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for the task: (returnval){ [ 1046.185133] env[61991]: value = "task-1130370" [ 1046.185133] env[61991]: _type = "Task" [ 1046.185133] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.190307] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130369, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.196170] env[61991]: DEBUG oslo_vmware.api [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130370, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.216240] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130367, 'name': ReconfigVM_Task, 'duration_secs': 0.198929} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.216585] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1046.217445] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3978f819-21b6-4340-a847-796c0886151d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.239927] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-f529fecd-03fc-4258-a8dc-f770503a03ee/volume-f529fecd-03fc-4258-a8dc-f770503a03ee.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.242919] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c111f5d-3953-49cf-9c78-7f1fe07e601c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.262653] env[61991]: DEBUG nova.compute.manager [req-fe7ea836-8ea7-4ebb-abff-4eebabb13bad req-c1d5b0c2-dad7-41ef-a1a4-f22bf88b79d8 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-vif-plugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1046.262875] env[61991]: DEBUG oslo_concurrency.lockutils [req-fe7ea836-8ea7-4ebb-abff-4eebabb13bad req-c1d5b0c2-dad7-41ef-a1a4-f22bf88b79d8 service nova] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.263112] env[61991]: DEBUG oslo_concurrency.lockutils [req-fe7ea836-8ea7-4ebb-abff-4eebabb13bad req-c1d5b0c2-dad7-41ef-a1a4-f22bf88b79d8 service nova] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.263388] env[61991]: DEBUG oslo_concurrency.lockutils [req-fe7ea836-8ea7-4ebb-abff-4eebabb13bad req-c1d5b0c2-dad7-41ef-a1a4-f22bf88b79d8 service nova] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.263458] env[61991]: DEBUG nova.compute.manager [req-fe7ea836-8ea7-4ebb-abff-4eebabb13bad req-c1d5b0c2-dad7-41ef-a1a4-f22bf88b79d8 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] No waiting events found dispatching network-vif-plugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.263620] env[61991]: WARNING nova.compute.manager [req-fe7ea836-8ea7-4ebb-abff-4eebabb13bad req-c1d5b0c2-dad7-41ef-a1a4-f22bf88b79d8 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received unexpected event network-vif-plugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 for instance with vm_state building and task_state spawning. [ 1046.265397] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1046.265397] env[61991]: value = "task-1130371" [ 1046.265397] env[61991]: _type = "Task" [ 1046.265397] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.274749] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130371, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.343513] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8801ad72-5b95-4d7d-8366-705c99bf7d4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.353329] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a8b154-ca27-49f1-b3ed-7deb69bb56f5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.386189] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519f4ce4-e75d-4694-bf64-b4b06b83102e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.394275] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c7c369-d0cd-43bf-8637-02404490a696 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.409316] env[61991]: DEBUG nova.compute.provider_tree [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.474332] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.474784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.597574] env[61991]: DEBUG nova.network.neutron [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Successfully updated port: 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.690114] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130369, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.694546] env[61991]: DEBUG oslo_vmware.api [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Task: {'id': task-1130370, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183971} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.694788] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.694978] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.695176] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.695356] env[61991]: INFO nova.compute.manager [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1046.695594] env[61991]: DEBUG oslo.service.loopingcall [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.695781] env[61991]: DEBUG nova.compute.manager [-] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.695877] env[61991]: DEBUG nova.network.neutron [-] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1046.774996] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130371, 'name': ReconfigVM_Task, 'duration_secs': 0.29905} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.775301] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-f529fecd-03fc-4258-a8dc-f770503a03ee/volume-f529fecd-03fc-4258-a8dc-f770503a03ee.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.775588] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1046.912584] env[61991]: DEBUG nova.scheduler.client.report [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.977893] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1047.103501] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.103501] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.103501] env[61991]: DEBUG nova.network.neutron [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.190013] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130369, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.283028] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af27cf0-4b97-4a9e-8c7a-79afede09a4f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.302095] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a359bc6-b150-43c3-84e6-219634554328 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.321325] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1047.400055] env[61991]: DEBUG nova.network.neutron [-] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.417661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.439466] env[61991]: INFO nova.scheduler.client.report [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Deleted allocations for instance 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b [ 1047.496015] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.496308] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.497931] env[61991]: INFO nova.compute.claims [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.637673] env[61991]: DEBUG nova.network.neutron [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1047.689594] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130369, 'name': ReconfigVM_Task, 'duration_secs': 1.462969} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.689911] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9/volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.697051] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bed7329a-c319-49ba-a048-6a5ce7c887c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.711078] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1047.711078] env[61991]: value = "task-1130372" [ 1047.711078] env[61991]: _type = "Task" [ 1047.711078] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.719688] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.819658] env[61991]: DEBUG nova.network.neutron [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.901378] env[61991]: INFO nova.compute.manager [-] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Took 1.21 seconds to deallocate network for instance. [ 1047.948027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4f23245b-d756-414c-bf6a-2d66925c8597 tempest-ServerRescueNegativeTestJSON-1892275759 tempest-ServerRescueNegativeTestJSON-1892275759-project-member] Lock "56e0ec0a-61d5-4394-a8f4-1a52ab61e08b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.304s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.221356] env[61991]: DEBUG oslo_vmware.api [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130372, 'name': ReconfigVM_Task, 'duration_secs': 0.229545} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.221601] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246969', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'name': 'volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be79140-21f9-493c-a840-2d905b8d66bf', 'attached_at': '', 'detached_at': '', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'serial': '69b6847b-fc53-4491-a7c8-08ee506cb1f9'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1048.292460] env[61991]: DEBUG nova.compute.manager [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.292639] env[61991]: DEBUG nova.compute.manager [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing instance network info cache due to event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1048.292842] env[61991]: DEBUG oslo_concurrency.lockutils [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.321992] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.322441] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance network_info: |[{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1048.322832] env[61991]: DEBUG oslo_concurrency.lockutils [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.323105] env[61991]: DEBUG nova.network.neutron [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.324706] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:36:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d2cca63-6f27-4e38-a6b1-6e0eda1401e6', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1048.337173] env[61991]: DEBUG oslo.service.loopingcall [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.338204] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1048.339156] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4335b89b-9a71-4119-bd4c-35efaa77a852 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.368303] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1048.368303] env[61991]: value = "task-1130373" [ 1048.368303] env[61991]: _type = "Task" [ 1048.368303] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.376315] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130373, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.409220] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.640753] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4fa386-d788-4a58-b142-41ae327efdcc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.648054] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bca33d-4eb9-4806-b443-9af80b7d1c1d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.678099] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91882e3c-f2fb-4890-957e-c71e1d33ede6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.685412] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e476bf53-8964-4629-b437-284d89d07504 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.699510] env[61991]: DEBUG nova.compute.provider_tree [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.886977] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130373, 'name': CreateVM_Task, 'duration_secs': 0.34056} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.887205] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.888399] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.888824] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.889252] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.889678] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4da3a6af-a3f1-44aa-8d26-804928564890 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.896186] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1048.896186] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba9de9-5cec-7ebe-7050-b6fe5e728fe1" [ 1048.896186] env[61991]: _type = "Task" [ 1048.896186] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.906519] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba9de9-5cec-7ebe-7050-b6fe5e728fe1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.007591] env[61991]: DEBUG nova.network.neutron [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Port 7277dc2e-f381-41e1-95de-31eda0f6cb27 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1049.101698] env[61991]: DEBUG nova.network.neutron [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updated VIF entry in instance network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.102644] env[61991]: DEBUG nova.network.neutron [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.203357] env[61991]: DEBUG nova.scheduler.client.report [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.263136] env[61991]: DEBUG nova.objects.instance [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'flavor' on Instance uuid 5be79140-21f9-493c-a840-2d905b8d66bf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.407532] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ba9de9-5cec-7ebe-7050-b6fe5e728fe1, 'name': SearchDatastore_Task, 'duration_secs': 0.009758} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.408389] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.408788] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1049.409163] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.409530] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.409973] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1049.410548] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e689ad8-8972-469b-857e-1d0fec5e25e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.419243] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1049.419584] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1049.422018] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7286d92f-57f3-4b4d-a5b5-30c7d0d9f87f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.426089] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1049.426089] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ff1cc7-a97a-2b3a-b897-64373a164455" [ 1049.426089] env[61991]: _type = "Task" [ 1049.426089] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.434315] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ff1cc7-a97a-2b3a-b897-64373a164455, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.605401] env[61991]: DEBUG oslo_concurrency.lockutils [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.605401] env[61991]: DEBUG nova.compute.manager [req-392c787d-c483-423a-a742-01192db653ee req-227bc3cd-bc49-488c-b802-4207a5a35c0d service nova] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Received event network-vif-deleted-3a2b9da3-3552-419b-a28e-ed7f76eb9607 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.711022] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.212s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.711022] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1049.712212] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.303s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.712627] env[61991]: DEBUG nova.objects.instance [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lazy-loading 'resources' on Instance uuid 81736f13-fc9e-4246-9d86-5a5bd7e15832 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.771748] env[61991]: DEBUG oslo_concurrency.lockutils [None req-706aef30-e915-4d5a-b7a0-a23214cbca03 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.277s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.937436] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52ff1cc7-a97a-2b3a-b897-64373a164455, 'name': SearchDatastore_Task, 'duration_secs': 0.008165} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.938273] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba673641-6838-4dab-901d-2eb3511f38b5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.943289] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1049.943289] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5288e524-2ba2-0205-7723-ba3414636ae1" [ 1049.943289] env[61991]: _type = "Task" [ 1049.943289] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.951033] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5288e524-2ba2-0205-7723-ba3414636ae1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.030750] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.030983] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.031218] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.216215] env[61991]: DEBUG nova.compute.utils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1050.220667] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1050.221324] env[61991]: DEBUG nova.network.neutron [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1050.264748] env[61991]: DEBUG nova.policy [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1050.269874] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.270130] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.340079] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed18096b-a4cf-425c-87c1-1d82ddd96b24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.347941] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1d51b3-77e8-44f6-bcbe-278ff1aa86dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.378275] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed060fd4-0809-4e4b-b91d-dc2b463d1447 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.385500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b33d335-6001-4165-a6e0-fc6e8b8e8774 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.398778] env[61991]: DEBUG nova.compute.provider_tree [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.452551] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5288e524-2ba2-0205-7723-ba3414636ae1, 'name': SearchDatastore_Task, 'duration_secs': 0.008606} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.452806] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.453069] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1050.453318] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72638572-512c-4bcd-8e87-9fb7ac6b89ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.460094] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1050.460094] env[61991]: value = "task-1130374" [ 1050.460094] env[61991]: _type = "Task" [ 1050.460094] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.467950] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130374, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.551246] env[61991]: DEBUG nova.network.neutron [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Successfully created port: 2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1050.721964] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1050.774775] env[61991]: INFO nova.compute.manager [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Detaching volume 69b6847b-fc53-4491-a7c8-08ee506cb1f9 [ 1050.815156] env[61991]: INFO nova.virt.block_device [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Attempting to driver detach volume 69b6847b-fc53-4491-a7c8-08ee506cb1f9 from mountpoint /dev/sdb [ 1050.815429] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1050.815623] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246969', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'name': 'volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be79140-21f9-493c-a840-2d905b8d66bf', 'attached_at': '', 'detached_at': '', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'serial': '69b6847b-fc53-4491-a7c8-08ee506cb1f9'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1050.816543] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46947b5b-a008-49a6-b6ff-4df4032da813 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.840920] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7d45a6-9ccc-4663-95fb-127a397c7e89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.848465] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c894ce-e8a6-4789-b3fb-ccc3ff1cc845 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.869901] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f376c4c-27b1-42c8-8496-5107868cc9db {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.885061] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] The volume has not been displaced from its original location: [datastore1] volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9/volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1050.890263] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1050.890594] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81dff719-9d30-44bc-9a96-4b4e33e589a2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.903687] env[61991]: DEBUG nova.scheduler.client.report [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.912319] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1050.912319] env[61991]: value = "task-1130375" [ 1050.912319] env[61991]: _type = "Task" [ 1050.912319] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.920449] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.969165] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130374, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43256} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.969455] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1050.969683] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.970036] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5121cba0-cb8a-4f29-9203-f6aeb14b85ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.975843] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1050.975843] env[61991]: value = "task-1130376" [ 1050.975843] env[61991]: _type = "Task" [ 1050.975843] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.983403] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130376, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.075546] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.075763] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.075950] env[61991]: DEBUG nova.network.neutron [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1051.409260] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.420715] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130375, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.432969] env[61991]: INFO nova.scheduler.client.report [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Deleted allocations for instance 81736f13-fc9e-4246-9d86-5a5bd7e15832 [ 1051.489945] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130376, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056947} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.489945] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.489945] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbde08a-5618-4ba9-890c-510437c17893 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.510197] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.510462] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-891b030b-0e4b-4a3b-836f-ba24e0f390e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.530618] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1051.530618] env[61991]: value = "task-1130377" [ 1051.530618] env[61991]: _type = "Task" [ 1051.530618] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.538842] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.732477] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1051.759107] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1051.759472] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1051.759673] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1051.759873] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1051.760032] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1051.760185] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1051.760405] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1051.760608] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1051.760787] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1051.760958] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1051.761150] env[61991]: DEBUG nova.virt.hardware [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.762057] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573be637-5393-4ede-a543-2884826e26c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.771751] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540745a2-a5d1-491e-b4ff-4101156011bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.799519] env[61991]: DEBUG nova.network.neutron [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [{"id": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "address": "fa:16:3e:37:4b:7d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7277dc2e-f3", "ovs_interfaceid": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.923593] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130375, 'name': ReconfigVM_Task, 'duration_secs': 0.512811} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.924725] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1051.929436] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22bfee21-3477-4a85-8b7f-3d8dc175bc61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.942972] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bbd01917-1ccc-4199-966f-95d0d40ece50 tempest-ServerDiskConfigTestJSON-664184996 tempest-ServerDiskConfigTestJSON-664184996-project-member] Lock "81736f13-fc9e-4246-9d86-5a5bd7e15832" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.383s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.948571] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1051.948571] env[61991]: value = "task-1130378" [ 1051.948571] env[61991]: _type = "Task" [ 1051.948571] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.956896] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130378, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.963117] env[61991]: DEBUG nova.compute.manager [req-73437dac-ec1d-4755-9374-b427280e51f2 req-5d796947-91a0-4821-b165-c8d7dcc3dc9d service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Received event network-vif-plugged-2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.963407] env[61991]: DEBUG oslo_concurrency.lockutils [req-73437dac-ec1d-4755-9374-b427280e51f2 req-5d796947-91a0-4821-b165-c8d7dcc3dc9d service nova] Acquiring lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.963723] env[61991]: DEBUG oslo_concurrency.lockutils [req-73437dac-ec1d-4755-9374-b427280e51f2 req-5d796947-91a0-4821-b165-c8d7dcc3dc9d service nova] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.963998] env[61991]: DEBUG oslo_concurrency.lockutils [req-73437dac-ec1d-4755-9374-b427280e51f2 req-5d796947-91a0-4821-b165-c8d7dcc3dc9d service nova] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.964250] env[61991]: DEBUG nova.compute.manager [req-73437dac-ec1d-4755-9374-b427280e51f2 req-5d796947-91a0-4821-b165-c8d7dcc3dc9d service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] No waiting events found dispatching network-vif-plugged-2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1051.964543] env[61991]: WARNING nova.compute.manager [req-73437dac-ec1d-4755-9374-b427280e51f2 req-5d796947-91a0-4821-b165-c8d7dcc3dc9d service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Received unexpected event network-vif-plugged-2b1cc073-3039-46af-8334-90cf1cc5c803 for instance with vm_state building and task_state spawning. [ 1052.040438] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130377, 'name': ReconfigVM_Task, 'duration_secs': 0.26895} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.040727] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.041636] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f010acb2-f9e7-47e3-8cbb-89ece7bc8c74 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.044213] env[61991]: DEBUG nova.network.neutron [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Successfully updated port: 2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1052.048248] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1052.048248] env[61991]: value = "task-1130379" [ 1052.048248] env[61991]: _type = "Task" [ 1052.048248] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.056265] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130379, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.302036] env[61991]: DEBUG oslo_concurrency.lockutils [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.458189] env[61991]: DEBUG oslo_vmware.api [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130378, 'name': ReconfigVM_Task, 'duration_secs': 0.170577} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.458513] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246969', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'name': 'volume-69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5be79140-21f9-493c-a840-2d905b8d66bf', 'attached_at': '', 'detached_at': '', 'volume_id': '69b6847b-fc53-4491-a7c8-08ee506cb1f9', 'serial': '69b6847b-fc53-4491-a7c8-08ee506cb1f9'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1052.552821] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-ea9e4210-9a4a-40db-9cc4-a59e3306019b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.552969] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-ea9e4210-9a4a-40db-9cc4-a59e3306019b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.553135] env[61991]: DEBUG nova.network.neutron [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1052.559875] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130379, 'name': Rename_Task, 'duration_secs': 0.153423} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.560167] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1052.560428] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f9e0821-7d2f-47e2-a971-828161a89758 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.567577] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1052.567577] env[61991]: value = "task-1130380" [ 1052.567577] env[61991]: _type = "Task" [ 1052.567577] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.583585] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130380, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.812979] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dcc60a-3c2e-43a6-a916-e4af69d9f460 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.821371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b3e891-ebd6-4a37-b528-7cce088d53b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.001664] env[61991]: DEBUG nova.objects.instance [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'flavor' on Instance uuid 5be79140-21f9-493c-a840-2d905b8d66bf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.077139] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130380, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.098749] env[61991]: DEBUG nova.network.neutron [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1053.283970] env[61991]: DEBUG nova.network.neutron [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Updating instance_info_cache with network_info: [{"id": "2b1cc073-3039-46af-8334-90cf1cc5c803", "address": "fa:16:3e:ee:77:d4", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1cc073-30", "ovs_interfaceid": "2b1cc073-3039-46af-8334-90cf1cc5c803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.579567] env[61991]: DEBUG oslo_vmware.api [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130380, 'name': PowerOnVM_Task, 'duration_secs': 0.649466} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.580368] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1053.580368] env[61991]: INFO nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1053.580368] env[61991]: DEBUG nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.581051] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a236992b-4759-47d9-952f-b75b6dcae282 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.786577] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-ea9e4210-9a4a-40db-9cc4-a59e3306019b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.787066] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Instance network_info: |[{"id": "2b1cc073-3039-46af-8334-90cf1cc5c803", "address": "fa:16:3e:ee:77:d4", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1cc073-30", "ovs_interfaceid": "2b1cc073-3039-46af-8334-90cf1cc5c803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1053.787533] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:77:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b1cc073-3039-46af-8334-90cf1cc5c803', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.795772] env[61991]: DEBUG oslo.service.loopingcall [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.795772] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.797619] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d187d80c-77af-4b80-96b8-188da0001437 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.819485] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.819485] env[61991]: value = "task-1130381" [ 1053.819485] env[61991]: _type = "Task" [ 1053.819485] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.827257] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130381, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.936392] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef3fbf7-7eb9-4ac2-acdc-b05e484e3d08 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.961826] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbf2137-51dd-4abe-96c9-d46a8b6a985a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.970722] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.994097] env[61991]: DEBUG nova.compute.manager [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Received event network-changed-2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.994097] env[61991]: DEBUG nova.compute.manager [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Refreshing instance network info cache due to event network-changed-2b1cc073-3039-46af-8334-90cf1cc5c803. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.994097] env[61991]: DEBUG oslo_concurrency.lockutils [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] Acquiring lock "refresh_cache-ea9e4210-9a4a-40db-9cc4-a59e3306019b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.994097] env[61991]: DEBUG oslo_concurrency.lockutils [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] Acquired lock "refresh_cache-ea9e4210-9a4a-40db-9cc4-a59e3306019b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.994097] env[61991]: DEBUG nova.network.neutron [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Refreshing network info cache for port 2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1054.010289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f19c36d0-ca2a-4112-af9a-44fbd6e27ece tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.740s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.100024] env[61991]: INFO nova.compute.manager [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Took 14.09 seconds to build instance. [ 1054.330944] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130381, 'name': CreateVM_Task, 'duration_secs': 0.2942} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.331161] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1054.331906] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.332118] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.332409] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.332671] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bfbaa7b-fdd4-4b74-b97d-9a4034af4595 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.337465] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1054.337465] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52514026-943d-fce5-68bd-7f1f41143ae9" [ 1054.337465] env[61991]: _type = "Task" [ 1054.337465] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.344806] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52514026-943d-fce5-68bd-7f1f41143ae9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.480586] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.480899] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2fd0c25-3cc9-4d36-bebb-9e8372f4bc95 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.490355] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1054.490355] env[61991]: value = "task-1130382" [ 1054.490355] env[61991]: _type = "Task" [ 1054.490355] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.502303] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.571358] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.571637] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.571848] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "5be79140-21f9-493c-a840-2d905b8d66bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.572072] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.572270] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.574891] env[61991]: INFO nova.compute.manager [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Terminating instance [ 1054.576691] env[61991]: DEBUG nova.compute.manager [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1054.576951] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.577759] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1039d15-128c-409d-8d5f-bd151568097c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.585591] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.585847] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f26d4f8f-f08c-4f80-96b8-2859cf06a520 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.592345] env[61991]: DEBUG oslo_vmware.api [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1054.592345] env[61991]: value = "task-1130383" [ 1054.592345] env[61991]: _type = "Task" [ 1054.592345] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.601467] env[61991]: DEBUG oslo_concurrency.lockutils [None req-7c9b8ba2-f60c-4683-8bf1-5c00b09f214b tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.603s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.601898] env[61991]: DEBUG oslo_vmware.api [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.769109] env[61991]: DEBUG nova.network.neutron [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Updated VIF entry in instance network info cache for port 2b1cc073-3039-46af-8334-90cf1cc5c803. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1054.769613] env[61991]: DEBUG nova.network.neutron [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Updating instance_info_cache with network_info: [{"id": "2b1cc073-3039-46af-8334-90cf1cc5c803", "address": "fa:16:3e:ee:77:d4", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b1cc073-30", "ovs_interfaceid": "2b1cc073-3039-46af-8334-90cf1cc5c803", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.853051] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52514026-943d-fce5-68bd-7f1f41143ae9, 'name': SearchDatastore_Task, 'duration_secs': 0.010371} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.856282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.856282] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.856282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.856282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.856282] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.856705] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b6e42b3-68c5-4990-8d7d-83ee653d7e22 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.867084] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.867713] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.868571] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33a4cc15-6eb6-4799-b4c3-e82706f122ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.883042] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1054.883042] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524fafec-d431-49ea-257d-bbcc9c228cc7" [ 1054.883042] env[61991]: _type = "Task" [ 1054.883042] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.891957] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524fafec-d431-49ea-257d-bbcc9c228cc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.000739] env[61991]: DEBUG oslo_vmware.api [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130382, 'name': PowerOnVM_Task, 'duration_secs': 0.411155} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.001191] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.001506] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-46a1b321-c35f-4acd-a87c-5605c040df07 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance 'c43df224-7c35-4b82-ba13-50f3a6f93f2e' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1055.103641] env[61991]: DEBUG oslo_vmware.api [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130383, 'name': PowerOffVM_Task, 'duration_secs': 0.293249} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.104092] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.104376] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1055.104736] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d2043e1-7b7d-4208-88a7-d19ab7b24249 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.166287] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1055.166671] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1055.166950] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleting the datastore file [datastore2] 5be79140-21f9-493c-a840-2d905b8d66bf {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1055.167278] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f802470-543c-4664-8f52-5c391da4b5eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.175037] env[61991]: DEBUG oslo_vmware.api [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for the task: (returnval){ [ 1055.175037] env[61991]: value = "task-1130385" [ 1055.175037] env[61991]: _type = "Task" [ 1055.175037] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.182453] env[61991]: DEBUG oslo_vmware.api [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.226485] env[61991]: DEBUG nova.compute.manager [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.226633] env[61991]: DEBUG nova.compute.manager [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing instance network info cache due to event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.226831] env[61991]: DEBUG oslo_concurrency.lockutils [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.226996] env[61991]: DEBUG oslo_concurrency.lockutils [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.227215] env[61991]: DEBUG nova.network.neutron [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1055.272451] env[61991]: DEBUG oslo_concurrency.lockutils [req-dfd1bd31-d821-46cf-81bc-fee285ba6dfd req-94910443-8f2c-4216-b3cc-b1752bc745bd service nova] Releasing lock "refresh_cache-ea9e4210-9a4a-40db-9cc4-a59e3306019b" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.393356] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]524fafec-d431-49ea-257d-bbcc9c228cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.01584} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.394174] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee24dfaf-470b-4d5b-8ed7-f138015d9bbf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.399702] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1055.399702] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525a7854-2be1-dc8c-5dd1-9ae172227a94" [ 1055.399702] env[61991]: _type = "Task" [ 1055.399702] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.406923] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525a7854-2be1-dc8c-5dd1-9ae172227a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.685524] env[61991]: DEBUG oslo_vmware.api [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Task: {'id': task-1130385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.509969} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.686028] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.686028] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.686146] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.686341] env[61991]: INFO nova.compute.manager [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1055.686588] env[61991]: DEBUG oslo.service.loopingcall [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.687071] env[61991]: DEBUG nova.compute.manager [-] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1055.687071] env[61991]: DEBUG nova.network.neutron [-] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1055.913040] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525a7854-2be1-dc8c-5dd1-9ae172227a94, 'name': SearchDatastore_Task, 'duration_secs': 0.009963} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.913483] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.913483] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] ea9e4210-9a4a-40db-9cc4-a59e3306019b/ea9e4210-9a4a-40db-9cc4-a59e3306019b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1055.913775] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9694558-cce1-4cbe-a99b-efa1c04ed955 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.924835] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1055.924835] env[61991]: value = "task-1130386" [ 1055.924835] env[61991]: _type = "Task" [ 1055.924835] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.932986] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.969815] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1055.970186] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1055.970365] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1056.194340] env[61991]: DEBUG nova.network.neutron [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updated VIF entry in instance network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1056.194742] env[61991]: DEBUG nova.network.neutron [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.363201] env[61991]: DEBUG nova.compute.manager [req-e14f4fd6-e264-4fb4-a78b-a7c0a1c46057 req-619dccb5-3f21-44ea-8cb7-676c3ea563e6 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Received event network-vif-deleted-f5641235-5202-4d60-a351-b93be008988e {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.363201] env[61991]: INFO nova.compute.manager [req-e14f4fd6-e264-4fb4-a78b-a7c0a1c46057 req-619dccb5-3f21-44ea-8cb7-676c3ea563e6 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Neutron deleted interface f5641235-5202-4d60-a351-b93be008988e; detaching it from the instance and deleting it from the info cache [ 1056.363201] env[61991]: DEBUG nova.network.neutron [req-e14f4fd6-e264-4fb4-a78b-a7c0a1c46057 req-619dccb5-3f21-44ea-8cb7-676c3ea563e6 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.440559] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130386, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.697684] env[61991]: DEBUG oslo_concurrency.lockutils [req-086ed1a7-5bff-48ea-9ff8-bec3c533438a req-6bf76b06-bb54-432b-be5a-43609edee8a6 service nova] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.832098] env[61991]: DEBUG nova.network.neutron [-] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.866035] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88b69a2d-a45c-4e38-9526-df5976635d56 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.874945] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4aa2140-388f-4f97-b484-6a5549f4f603 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.912355] env[61991]: DEBUG nova.compute.manager [req-e14f4fd6-e264-4fb4-a78b-a7c0a1c46057 req-619dccb5-3f21-44ea-8cb7-676c3ea563e6 service nova] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Detach interface failed, port_id=f5641235-5202-4d60-a351-b93be008988e, reason: Instance 5be79140-21f9-493c-a840-2d905b8d66bf could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1056.935009] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130386, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5451} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.935314] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] ea9e4210-9a4a-40db-9cc4-a59e3306019b/ea9e4210-9a4a-40db-9cc4-a59e3306019b.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.935520] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.935774] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb952df9-9f1c-4eb8-abb0-203bafa4b2a7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.942069] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1056.942069] env[61991]: value = "task-1130387" [ 1056.942069] env[61991]: _type = "Task" [ 1056.942069] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.949176] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.337696] env[61991]: INFO nova.compute.manager [-] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Took 1.65 seconds to deallocate network for instance. [ 1057.451445] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063892} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.451909] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.452486] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813460e1-3431-4730-8e0f-85118f6e0403 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.475352] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] ea9e4210-9a4a-40db-9cc4-a59e3306019b/ea9e4210-9a4a-40db-9cc4-a59e3306019b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.475649] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-518157f8-88fb-4302-b448-7b614c147020 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.495689] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1057.495689] env[61991]: value = "task-1130388" [ 1057.495689] env[61991]: _type = "Task" [ 1057.495689] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.503852] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130388, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.820984] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.821304] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.821530] env[61991]: DEBUG nova.compute.manager [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Going to confirm migration 8 {{(pid=61991) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1057.843920] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.844257] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.844529] env[61991]: DEBUG nova.objects.instance [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lazy-loading 'resources' on Instance uuid 5be79140-21f9-493c-a840-2d905b8d66bf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.005954] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130388, 'name': ReconfigVM_Task, 'duration_secs': 0.499243} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.005954] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Reconfigured VM instance instance-0000006d to attach disk [datastore2] ea9e4210-9a4a-40db-9cc4-a59e3306019b/ea9e4210-9a4a-40db-9cc4-a59e3306019b.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.007198] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fbf48fef-94fa-4ef6-9800-51bbe4d4d41c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.013355] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1058.013355] env[61991]: value = "task-1130389" [ 1058.013355] env[61991]: _type = "Task" [ 1058.013355] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.021655] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130389, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.386600] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.386829] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquired lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.387043] env[61991]: DEBUG nova.network.neutron [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1058.387245] env[61991]: DEBUG nova.objects.instance [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'info_cache' on Instance uuid c43df224-7c35-4b82-ba13-50f3a6f93f2e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.471784] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6124e88-fb33-4d95-966b-d835af56b306 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.480420] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887bfbc4-8831-4564-aa3c-99ac99decc04 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.516615] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3236fc8-0970-4303-8930-113f4034607e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.530086] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162ee3e3-d73f-47fe-84ee-a2921f23e3a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.534210] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130389, 'name': Rename_Task, 'duration_secs': 0.120901} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.534947] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1058.535587] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-506a2a61-2dae-4eed-a94a-7bbd16362ff5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.545254] env[61991]: DEBUG nova.compute.provider_tree [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.550165] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.550165] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.550165] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1058.551962] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1058.551962] env[61991]: value = "task-1130390" [ 1058.551962] env[61991]: _type = "Task" [ 1058.551962] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.561114] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.049142] env[61991]: DEBUG nova.scheduler.client.report [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.064142] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130390, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.558701] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.568189] env[61991]: DEBUG oslo_vmware.api [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130390, 'name': PowerOnVM_Task, 'duration_secs': 0.526886} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.568490] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.568709] env[61991]: INFO nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Took 7.84 seconds to spawn the instance on the hypervisor. [ 1059.569046] env[61991]: DEBUG nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1059.569718] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3b0ab3-fecc-4bee-8e9f-df1e5fd87892 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.653083] env[61991]: DEBUG nova.network.neutron [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [{"id": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "address": "fa:16:3e:37:4b:7d", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7277dc2e-f3", "ovs_interfaceid": "7277dc2e-f381-41e1-95de-31eda0f6cb27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.718146] env[61991]: INFO nova.scheduler.client.report [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Deleted allocations for instance 5be79140-21f9-493c-a840-2d905b8d66bf [ 1059.807288] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updating instance_info_cache with network_info: [{"id": "96d11c6e-409f-4baf-983b-c066592aa09b", "address": "fa:16:3e:76:45:34", "network": {"id": "d09176cd-3927-4f3e-9d2d-72e2880d1617", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1447768797-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adbd4ea9b0744f28a0d4a46b4a04d683", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dced2f3d-7fd3-4a42-836d-9f02dab4c949", "external-id": "nsx-vlan-transportzone-117", "segmentation_id": 117, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96d11c6e-40", "ovs_interfaceid": "96d11c6e-409f-4baf-983b-c066592aa09b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.087214] env[61991]: INFO nova.compute.manager [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Took 12.60 seconds to build instance. [ 1060.156243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Releasing lock "refresh_cache-c43df224-7c35-4b82-ba13-50f3a6f93f2e" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.156549] env[61991]: DEBUG nova.objects.instance [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'migration_context' on Instance uuid c43df224-7c35-4b82-ba13-50f3a6f93f2e {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.225744] env[61991]: DEBUG oslo_concurrency.lockutils [None req-996c659e-cd7e-47e5-9b5f-d8ab62cf0c61 tempest-AttachVolumeNegativeTest-958507313 tempest-AttachVolumeNegativeTest-958507313-project-member] Lock "5be79140-21f9-493c-a840-2d905b8d66bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.654s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.309913] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-100d5fc0-e9d9-4892-9f60-99d657c533ed" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.310233] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1060.310376] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.310535] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.310683] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.310829] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.310969] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.311129] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.311265] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1060.311422] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.589538] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c0a2b286-6cae-4170-800c-3661ab96bfe9 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.115s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.659646] env[61991]: DEBUG nova.objects.base [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1060.660643] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16691e29-6e9a-44f8-a921-fdec67bb7163 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.679942] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b337bce-e63b-48ab-9c0c-ce80a7d6d6a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.685647] env[61991]: DEBUG oslo_vmware.api [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1060.685647] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5265428d-464b-669d-eea1-a07e5d7820ff" [ 1060.685647] env[61991]: _type = "Task" [ 1060.685647] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.693265] env[61991]: DEBUG oslo_vmware.api [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5265428d-464b-669d-eea1-a07e5d7820ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.814176] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.814416] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.814587] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.814740] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1060.815704] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fad9b30-b336-4f8d-95bc-8b2a9f7fa71a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.824346] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae39f575-b60b-4cde-8de8-8436cd2597a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.839490] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c7af85-83b9-45de-bef4-ce46c23e26e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.846546] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0c1f8a-86ca-4734-bf2d-f17de23f244f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.876240] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180254MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1060.876450] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.876691] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.196298] env[61991]: DEBUG oslo_vmware.api [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5265428d-464b-669d-eea1-a07e5d7820ff, 'name': SearchDatastore_Task, 'duration_secs': 0.007199} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.196656] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.234316] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.234611] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.234833] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.235033] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.235213] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.237513] env[61991]: INFO nova.compute.manager [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Terminating instance [ 1061.239330] env[61991]: DEBUG nova.compute.manager [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.239527] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.240341] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0df592a-e1b1-48ac-806b-3ec4759a61fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.247528] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.247743] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93d57ff6-964c-4008-a88b-d1c66cba5a9e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.252938] env[61991]: DEBUG oslo_vmware.api [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1061.252938] env[61991]: value = "task-1130392" [ 1061.252938] env[61991]: _type = "Task" [ 1061.252938] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.260412] env[61991]: DEBUG oslo_vmware.api [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130392, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.762422] env[61991]: DEBUG oslo_vmware.api [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130392, 'name': PowerOffVM_Task, 'duration_secs': 0.199746} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.762692] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.762866] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.763144] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0796e33d-a217-4c72-ab06-4e1f6525e236 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.823897] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.824196] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.824435] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore2] ea9e4210-9a4a-40db-9cc4-a59e3306019b {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.824705] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4018017-7529-401f-918f-e6c0b129bece {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.831314] env[61991]: DEBUG oslo_vmware.api [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1061.831314] env[61991]: value = "task-1130394" [ 1061.831314] env[61991]: _type = "Task" [ 1061.831314] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.839312] env[61991]: DEBUG oslo_vmware.api [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130394, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.886491] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Applying migration context for instance c43df224-7c35-4b82-ba13-50f3a6f93f2e as it has an incoming, in-progress migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd. Migration status is confirming {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1061.887678] env[61991]: INFO nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating resource usage from migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd [ 1061.906220] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.906413] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 100d5fc0-e9d9-4892-9f60-99d657c533ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.906476] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance b26f02c6-a8ea-402d-922d-7879ccb28b92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.906561] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 84a0dee7-c33b-494a-ad38-83da6ab44ce2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.906747] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1061.906948] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance c43df224-7c35-4b82-ba13-50f3a6f93f2e actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.907154] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 22f59842-073b-4bca-bf4f-a83552a90582 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.907327] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance ea9e4210-9a4a-40db-9cc4-a59e3306019b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1061.907477] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1061.907613] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1062.008274] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e30d604-3b21-4e61-916b-e307bbc72a0b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.017579] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87164937-409b-41f5-a353-8423fffeaa48 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.047169] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ef5fec-c750-48ae-a712-94d58409eca5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.054507] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e88e8cc-0af7-496d-896e-f7861fabc181 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.067478] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.341067] env[61991]: DEBUG oslo_vmware.api [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130394, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187714} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.341413] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.341545] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.341667] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.341839] env[61991]: INFO nova.compute.manager [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1062.342101] env[61991]: DEBUG oslo.service.loopingcall [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.342296] env[61991]: DEBUG nova.compute.manager [-] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1062.342395] env[61991]: DEBUG nova.network.neutron [-] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1062.571078] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.816241] env[61991]: DEBUG nova.compute.manager [req-988016c1-f2c8-4907-bec6-67a7c5ac832a req-096fbb74-ec63-4abf-aae9-063263c7dc10 service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Received event network-vif-deleted-2b1cc073-3039-46af-8334-90cf1cc5c803 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.816358] env[61991]: INFO nova.compute.manager [req-988016c1-f2c8-4907-bec6-67a7c5ac832a req-096fbb74-ec63-4abf-aae9-063263c7dc10 service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Neutron deleted interface 2b1cc073-3039-46af-8334-90cf1cc5c803; detaching it from the instance and deleting it from the info cache [ 1062.816541] env[61991]: DEBUG nova.network.neutron [req-988016c1-f2c8-4907-bec6-67a7c5ac832a req-096fbb74-ec63-4abf-aae9-063263c7dc10 service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.075591] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1063.075802] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.199s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.076115] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.880s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.077445] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1063.077591] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Cleaning up deleted instances {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1063.288140] env[61991]: DEBUG nova.network.neutron [-] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.320011] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e75efd17-6ccd-4324-b8c5-e79d8edfe82f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.330032] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adf412f-871d-4350-a5e2-7bbeaaad9b84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.357561] env[61991]: DEBUG nova.compute.manager [req-988016c1-f2c8-4907-bec6-67a7c5ac832a req-096fbb74-ec63-4abf-aae9-063263c7dc10 service nova] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Detach interface failed, port_id=2b1cc073-3039-46af-8334-90cf1cc5c803, reason: Instance ea9e4210-9a4a-40db-9cc4-a59e3306019b could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1063.595423] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] There are 57 instances to clean {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1063.595605] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 81736f13-fc9e-4246-9d86-5a5bd7e15832] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.700808] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205bdb7e-14bc-4e75-afbf-ba9857806822 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.708440] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dea639-8b4c-4a46-9f9a-3272f5fda807 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.738174] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa067f7-e467-4ded-a7e6-0fa167ac3f8a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.745243] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d716c05-525f-43f3-baa8-0f5b101afe30 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.758675] env[61991]: DEBUG nova.compute.provider_tree [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.790819] env[61991]: INFO nova.compute.manager [-] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Took 1.45 seconds to deallocate network for instance. [ 1064.099848] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 39d9d434-e6d3-49e0-8a4c-adc8a8eb7af6] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.261581] env[61991]: DEBUG nova.scheduler.client.report [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.297703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.603047] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 5be79140-21f9-493c-a840-2d905b8d66bf] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.106035] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 679d3629-63e0-452f-9cee-1da4b9b843de] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.272042] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.196s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.274616] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.977s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.274887] env[61991]: DEBUG nova.objects.instance [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'resources' on Instance uuid ea9e4210-9a4a-40db-9cc4-a59e3306019b {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.608881] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 8005b8f6-6f4d-4f72-be5d-955fa114408d] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.832016] env[61991]: INFO nova.scheduler.client.report [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted allocation for migration 00c31364-8c4b-40cf-bd77-2e1d1128d9dd [ 1065.876849] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea144a4-b6a1-40ad-8888-bdd2bdb144d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.885267] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73220f3-f818-4f27-87c2-d3c25451d44e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.915475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d40878-6784-404c-a9ca-ba59d9d8d318 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.922266] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7114d465-52f7-455d-90c1-d89f540a9f7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.935267] env[61991]: DEBUG nova.compute.provider_tree [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.112526] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 892323a3-8661-47da-a52a-d7e0557ea300] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.339596] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f6bae3c3-1d26-4e38-ae7b-e22da09abb54 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.518s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.437891] env[61991]: DEBUG nova.scheduler.client.report [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.514832] env[61991]: INFO nova.compute.manager [None req-118710eb-9652-47de-9a28-d8dd469662f9 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Get console output [ 1066.515259] env[61991]: WARNING nova.virt.vmwareapi.driver [None req-118710eb-9652-47de-9a28-d8dd469662f9 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] The console log is missing. Check your VSPC configuration [ 1066.615478] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 7ad63ade-cff7-4a32-b567-783db726de55] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.942960] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.668s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.969265] env[61991]: INFO nova.scheduler.client.report [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance ea9e4210-9a4a-40db-9cc4-a59e3306019b [ 1067.118064] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 1730f9b3-69ad-476b-b3ab-3e1345f2a115] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.477563] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6343a26b-8b04-4716-acda-7644bf4c0a0c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "ea9e4210-9a4a-40db-9cc4-a59e3306019b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.242s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.620835] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 68268135-4cf4-4732-9104-1720bbb21acf] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.126471] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3b647920-4a69-4d1d-adb1-6fbf716e5514] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.236440] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "a0901469-3a2a-4d91-90fc-1909c0573caf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.236676] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.630308] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 586bf51a-d1c4-44e8-863b-b229a3ca4f2a] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.739822] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1069.133812] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 5364f93c-ca53-480b-8397-c331dfdeed3f] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.264137] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.264451] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.266627] env[61991]: INFO nova.compute.claims [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.637188] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 0745dd19-6b08-447a-8333-047aeaa1fa83] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.141082] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 077c5def-84cc-48f9-be90-410100666773] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.368118] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9f41c9-9a92-4f45-b3c2-ba7b839856ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.375881] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ac9081-3d85-47fa-9986-7d395bbc603a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.406107] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf858949-7580-407f-8c9e-d0eaf5464d9d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.412973] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fd298a-5779-448f-80fd-acfe7d0ff111 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.425692] env[61991]: DEBUG nova.compute.provider_tree [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.645091] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: a21cc5c6-abd8-45c6-b24b-33d3374df65c] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.928131] env[61991]: DEBUG nova.scheduler.client.report [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.148461] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 69a272a7-8da6-4fb2-95be-85cfb7d70d44] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.432439] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.432956] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1071.651578] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 902f158c-fcc4-40a2-abbd-491bdfd4b77a] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.938061] env[61991]: DEBUG nova.compute.utils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1071.939679] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1071.940094] env[61991]: DEBUG nova.network.neutron [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1071.989518] env[61991]: DEBUG nova.policy [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1072.154799] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 0dcf415e-b554-4c82-b127-9f0e22176031] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.250807] env[61991]: DEBUG nova.network.neutron [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Successfully created port: 10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1072.441048] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1072.658138] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 39004ce4-c96b-4911-9b64-3c508ddcecaf] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.161404] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: b3a2ea40-5c4c-4e7d-95b1-38a18c429cec] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.450120] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1073.476189] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.476481] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.476620] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.476795] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.477032] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.477240] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.477471] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.477627] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.477796] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.477965] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.478181] env[61991]: DEBUG nova.virt.hardware [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.479043] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436fc938-7065-4577-89e6-e85586842374 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.487095] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6f59f3-b046-4bac-9633-1ee87d268f3c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.664912] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f902f00f-eb2f-418a-bbfa-535442446681] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.713387] env[61991]: DEBUG nova.compute.manager [req-6e7e53a1-df49-4c3d-82ad-5054040b7cad req-ed09ad13-9505-4928-84ba-701bc27be30a service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Received event network-vif-plugged-10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.713387] env[61991]: DEBUG oslo_concurrency.lockutils [req-6e7e53a1-df49-4c3d-82ad-5054040b7cad req-ed09ad13-9505-4928-84ba-701bc27be30a service nova] Acquiring lock "a0901469-3a2a-4d91-90fc-1909c0573caf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.713387] env[61991]: DEBUG oslo_concurrency.lockutils [req-6e7e53a1-df49-4c3d-82ad-5054040b7cad req-ed09ad13-9505-4928-84ba-701bc27be30a service nova] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.713387] env[61991]: DEBUG oslo_concurrency.lockutils [req-6e7e53a1-df49-4c3d-82ad-5054040b7cad req-ed09ad13-9505-4928-84ba-701bc27be30a service nova] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.713520] env[61991]: DEBUG nova.compute.manager [req-6e7e53a1-df49-4c3d-82ad-5054040b7cad req-ed09ad13-9505-4928-84ba-701bc27be30a service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] No waiting events found dispatching network-vif-plugged-10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1073.713609] env[61991]: WARNING nova.compute.manager [req-6e7e53a1-df49-4c3d-82ad-5054040b7cad req-ed09ad13-9505-4928-84ba-701bc27be30a service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Received unexpected event network-vif-plugged-10412131-7e63-4541-bac9-21d6c54bc5e2 for instance with vm_state building and task_state spawning. [ 1074.168597] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 97913aea-48c1-4fda-aee3-578e89a08f71] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.177316] env[61991]: DEBUG nova.network.neutron [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Successfully updated port: 10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1074.201174] env[61991]: DEBUG nova.compute.manager [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Received event network-changed-10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1074.201401] env[61991]: DEBUG nova.compute.manager [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Refreshing instance network info cache due to event network-changed-10412131-7e63-4541-bac9-21d6c54bc5e2. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1074.201709] env[61991]: DEBUG oslo_concurrency.lockutils [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] Acquiring lock "refresh_cache-a0901469-3a2a-4d91-90fc-1909c0573caf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.201906] env[61991]: DEBUG oslo_concurrency.lockutils [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] Acquired lock "refresh_cache-a0901469-3a2a-4d91-90fc-1909c0573caf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.202132] env[61991]: DEBUG nova.network.neutron [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Refreshing network info cache for port 10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1074.673138] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: d2d9c3f0-f53a-48a7-a3d1-6765adc0dda3] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.680109] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-a0901469-3a2a-4d91-90fc-1909c0573caf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.734575] env[61991]: DEBUG nova.network.neutron [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1074.810632] env[61991]: DEBUG nova.network.neutron [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.176487] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9d1ac179-3d7e-4f0b-a80d-033eba447cab] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.314931] env[61991]: DEBUG oslo_concurrency.lockutils [req-3769a52c-f87d-4af4-9a33-a6094b384cf4 req-07b200da-77f6-4346-b174-98bf31b663ff service nova] Releasing lock "refresh_cache-a0901469-3a2a-4d91-90fc-1909c0573caf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.315337] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-a0901469-3a2a-4d91-90fc-1909c0573caf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.315495] env[61991]: DEBUG nova.network.neutron [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1075.679451] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 7634e00f-48ee-40a0-8625-2f8dc83cd58d] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.845308] env[61991]: DEBUG nova.network.neutron [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1075.975717] env[61991]: DEBUG nova.network.neutron [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Updating instance_info_cache with network_info: [{"id": "10412131-7e63-4541-bac9-21d6c54bc5e2", "address": "fa:16:3e:d4:8c:b0", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10412131-7e", "ovs_interfaceid": "10412131-7e63-4541-bac9-21d6c54bc5e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.183144] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f7eab1a9-2a8a-4664-8736-ca599e207c31] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.478783] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-a0901469-3a2a-4d91-90fc-1909c0573caf" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.479016] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Instance network_info: |[{"id": "10412131-7e63-4541-bac9-21d6c54bc5e2", "address": "fa:16:3e:d4:8c:b0", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10412131-7e", "ovs_interfaceid": "10412131-7e63-4541-bac9-21d6c54bc5e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1076.479557] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:8c:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10412131-7e63-4541-bac9-21d6c54bc5e2', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.487396] env[61991]: DEBUG oslo.service.loopingcall [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.487642] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.487908] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62d77236-ba0e-4bfb-bc3d-26a11ef5af7f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.509463] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.509463] env[61991]: value = "task-1130397" [ 1076.509463] env[61991]: _type = "Task" [ 1076.509463] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.686138] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 56e0ec0a-61d5-4394-a8f4-1a52ab61e08b] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.736355] env[61991]: DEBUG nova.compute.manager [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Stashing vm_state: active {{(pid=61991) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1077.019917] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130397, 'name': CreateVM_Task, 'duration_secs': 0.338963} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.020102] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.020763] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.020942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.021282] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1077.021523] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a9bcf79-f9a5-4399-87d7-3751e3e93ec8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.025555] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1077.025555] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526411f7-f72a-f845-757b-9b6966265a88" [ 1077.025555] env[61991]: _type = "Task" [ 1077.025555] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.032364] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526411f7-f72a-f845-757b-9b6966265a88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.189726] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3183190e-3202-40be-894c-e47b186708b5] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.255013] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.255335] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.537080] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526411f7-f72a-f845-757b-9b6966265a88, 'name': SearchDatastore_Task, 'duration_secs': 0.009063} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.537390] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.537623] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.537859] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.538052] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.538272] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.538522] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1baa8590-45cc-4049-a844-c8b11e4fffbf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.547497] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.547643] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.548361] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d07d580-dfd1-4ae1-9488-c478e649aa17 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.552951] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1077.552951] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d886b0-17e5-e846-6648-40516b28754e" [ 1077.552951] env[61991]: _type = "Task" [ 1077.552951] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.560177] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d886b0-17e5-e846-6648-40516b28754e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.693096] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 7a468ab7-6ca1-43e2-a9b2-89e7640d8148] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.759610] env[61991]: INFO nova.compute.claims [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.063462] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d886b0-17e5-e846-6648-40516b28754e, 'name': SearchDatastore_Task, 'duration_secs': 0.008308} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.064230] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35e24983-70fe-41ad-abc7-0274ef5625d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.069312] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1078.069312] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526206c5-e924-a337-911c-884f5bae3065" [ 1078.069312] env[61991]: _type = "Task" [ 1078.069312] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.076425] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526206c5-e924-a337-911c-884f5bae3065, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.196680] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 56188232-bed0-4a4b-a4bc-01edbb85cbe4] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.266115] env[61991]: INFO nova.compute.resource_tracker [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating resource usage from migration 5d92b469-b2f9-4e3d-9b16-dfe03939edfc [ 1078.355491] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecaad7c5-1a91-453e-a1cc-fcc996779376 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.362997] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597e832d-9910-46df-8a64-2cac7af535a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.391170] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b62d40-d22c-420a-86f8-a2077409ba18 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.398487] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192ac3b7-776c-4b8f-a08e-fccf2eb4c260 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.413159] env[61991]: DEBUG nova.compute.provider_tree [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.579993] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526206c5-e924-a337-911c-884f5bae3065, 'name': SearchDatastore_Task, 'duration_secs': 0.008787} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.580233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.580492] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] a0901469-3a2a-4d91-90fc-1909c0573caf/a0901469-3a2a-4d91-90fc-1909c0573caf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1078.580735] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cc2d9d7-6b86-4692-b555-3c08b31e3d6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.586819] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1078.586819] env[61991]: value = "task-1130398" [ 1078.586819] env[61991]: _type = "Task" [ 1078.586819] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.593997] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130398, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.700339] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c1d4d2eb-810c-42f9-a244-e5aed6e8d517] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.916689] env[61991]: DEBUG nova.scheduler.client.report [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.096765] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130398, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.418269} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.097074] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] a0901469-3a2a-4d91-90fc-1909c0573caf/a0901469-3a2a-4d91-90fc-1909c0573caf.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1079.097307] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1079.097543] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83c4c9e7-ccb8-4ce0-bde1-6b8480dc6894 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.103786] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1079.103786] env[61991]: value = "task-1130399" [ 1079.103786] env[61991]: _type = "Task" [ 1079.103786] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.110573] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130399, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.203833] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 96627657-d3d9-4a64-b83e-87323bd10c03] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.422367] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.167s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.422595] env[61991]: INFO nova.compute.manager [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Migrating [ 1079.613636] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130399, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055286} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.613910] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.614671] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6a368f-01e9-4eef-8d60-768e280c605a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.635511] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] a0901469-3a2a-4d91-90fc-1909c0573caf/a0901469-3a2a-4d91-90fc-1909c0573caf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.635760] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05aaf001-6b0a-4f63-9302-01676799a405 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.657175] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1079.657175] env[61991]: value = "task-1130400" [ 1079.657175] env[61991]: _type = "Task" [ 1079.657175] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.664614] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130400, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.707168] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 15215bba-8160-408b-9e8f-afbb9e71d50e] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.937555] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.937784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.937986] env[61991]: DEBUG nova.network.neutron [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1080.166549] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130400, 'name': ReconfigVM_Task, 'duration_secs': 0.264427} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.166863] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Reconfigured VM instance instance-0000006e to attach disk [datastore1] a0901469-3a2a-4d91-90fc-1909c0573caf/a0901469-3a2a-4d91-90fc-1909c0573caf.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.167524] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87661f2d-7d37-43de-b28f-8919e2880846 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.173832] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1080.173832] env[61991]: value = "task-1130401" [ 1080.173832] env[61991]: _type = "Task" [ 1080.173832] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.181173] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130401, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.210240] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: d36bd981-9ea3-46f6-8376-ac1e0c3bf61e] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.646310] env[61991]: DEBUG nova.network.neutron [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.683827] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130401, 'name': Rename_Task, 'duration_secs': 0.185202} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.684114] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.684358] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dd1df36-a9c0-4e8c-ba0b-2b357175e786 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.690886] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1080.690886] env[61991]: value = "task-1130402" [ 1080.690886] env[61991]: _type = "Task" [ 1080.690886] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.699051] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.713573] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 62b3e5ae-b90c-47f8-95c4-14587dbf647d] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.148980] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.201264] env[61991]: DEBUG oslo_vmware.api [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130402, 'name': PowerOnVM_Task, 'duration_secs': 0.417726} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.201549] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.201747] env[61991]: INFO nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Took 7.75 seconds to spawn the instance on the hypervisor. [ 1081.201930] env[61991]: DEBUG nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1081.202699] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b3517a-639d-4011-b636-ce212568eb90 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.216194] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 49324fef-ad48-451b-a5ce-d9a1231137db] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.719170] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f6945e87-3b17-4e0e-845f-3bacf0fbe5ac] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.720748] env[61991]: INFO nova.compute.manager [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Took 12.48 seconds to build instance. [ 1082.223048] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 884e465b-e14f-4114-81a6-40e2e8694558] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.224604] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c6b35a03-40b6-46bc-909d-ba3d7ad1d56f tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.988s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.651918] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "a0901469-3a2a-4d91-90fc-1909c0573caf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.652215] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.652441] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "a0901469-3a2a-4d91-90fc-1909c0573caf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.652630] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.652804] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.654917] env[61991]: INFO nova.compute.manager [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Terminating instance [ 1082.656625] env[61991]: DEBUG nova.compute.manager [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1082.656819] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1082.657692] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71254e61-1cac-49bb-9a4a-04a6b5a5cfbf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.663564] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f3ee2e-ccf6-4180-ac85-e6cd27323331 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.668190] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1082.668721] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af253ff5-65f0-48c3-9590-860692cdf76a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.682637] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 0 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1082.686760] env[61991]: DEBUG oslo_vmware.api [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1082.686760] env[61991]: value = "task-1130403" [ 1082.686760] env[61991]: _type = "Task" [ 1082.686760] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.695087] env[61991]: DEBUG oslo_vmware.api [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.725357] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3e62f219-7fec-470f-9d0a-e292d62e076c] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.189331] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.189674] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8516042e-ed46-4994-b4df-816a4331fb3d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.200331] env[61991]: DEBUG oslo_vmware.api [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130403, 'name': PowerOffVM_Task, 'duration_secs': 0.179788} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.202390] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1083.202390] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1083.202390] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1083.202390] env[61991]: value = "task-1130404" [ 1083.202390] env[61991]: _type = "Task" [ 1083.202390] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.202390] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c255c501-d5cf-4b2f-af9f-51bbc19b33ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.213150] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.228766] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: e3e5ec72-eec1-4155-bab2-038e3d5e38a8] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.272093] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1083.272407] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1083.272684] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore1] a0901469-3a2a-4d91-90fc-1909c0573caf {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1083.272994] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecf3c48d-1819-475e-be54-b2e54f3feb97 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.279994] env[61991]: DEBUG oslo_vmware.api [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1083.279994] env[61991]: value = "task-1130406" [ 1083.279994] env[61991]: _type = "Task" [ 1083.279994] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.288139] env[61991]: DEBUG oslo_vmware.api [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130406, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.714168] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130404, 'name': PowerOffVM_Task, 'duration_secs': 0.20267} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.714411] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1083.714602] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 17 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.732046] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c279a82f-bf4c-4a59-b04a-38d31070b0e5] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.790896] env[61991]: DEBUG oslo_vmware.api [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130406, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169304} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.791159] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1083.791352] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1083.791524] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1083.791695] env[61991]: INFO nova.compute.manager [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1083.791930] env[61991]: DEBUG oslo.service.loopingcall [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.792163] env[61991]: DEBUG nova.compute.manager [-] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1083.792305] env[61991]: DEBUG nova.network.neutron [-] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1084.032574] env[61991]: DEBUG nova.compute.manager [req-ec7fc490-4a0d-4194-8b0a-f502cf16921d req-2f5b776b-ff6c-4eeb-ba7e-d4b7fa438483 service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Received event network-vif-deleted-10412131-7e63-4541-bac9-21d6c54bc5e2 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.032770] env[61991]: INFO nova.compute.manager [req-ec7fc490-4a0d-4194-8b0a-f502cf16921d req-2f5b776b-ff6c-4eeb-ba7e-d4b7fa438483 service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Neutron deleted interface 10412131-7e63-4541-bac9-21d6c54bc5e2; detaching it from the instance and deleting it from the info cache [ 1084.032947] env[61991]: DEBUG nova.network.neutron [req-ec7fc490-4a0d-4194-8b0a-f502cf16921d req-2f5b776b-ff6c-4eeb-ba7e-d4b7fa438483 service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.221226] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.221466] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.221630] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.221817] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.221966] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.222185] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.222337] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.222836] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.222836] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.222836] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.223045] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.228077] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff7e7e4b-ec3b-422e-92cb-c46e9fedf002 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.238812] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 0aa67bb1-59d3-42e4-8f12-47e61a130dfc] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.246906] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1084.246906] env[61991]: value = "task-1130407" [ 1084.246906] env[61991]: _type = "Task" [ 1084.246906] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.254273] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.512900] env[61991]: DEBUG nova.network.neutron [-] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.535374] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a0c8a71-1473-401a-b474-f93a0c8d53b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.544895] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed4f588-3456-4080-8673-f7c37bef1de4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.571528] env[61991]: DEBUG nova.compute.manager [req-ec7fc490-4a0d-4194-8b0a-f502cf16921d req-2f5b776b-ff6c-4eeb-ba7e-d4b7fa438483 service nova] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Detach interface failed, port_id=10412131-7e63-4541-bac9-21d6c54bc5e2, reason: Instance a0901469-3a2a-4d91-90fc-1909c0573caf could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1084.741987] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: d6a2d226-5097-4461-b2ce-fa698f1066d5] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.756055] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130407, 'name': ReconfigVM_Task, 'duration_secs': 0.180368} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.756055] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 33 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1085.015077] env[61991]: INFO nova.compute.manager [-] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Took 1.22 seconds to deallocate network for instance. [ 1085.245245] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 671e0197-4f96-4660-be99-ea1d1c0588e1] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.261407] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.261662] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.261823] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.262015] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.262175] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.262330] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.262536] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.262700] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.262867] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.263041] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.263229] env[61991]: DEBUG nova.virt.hardware [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.268570] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1085.269144] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1263b89c-2e7f-40d4-b46a-513393016dc4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.288525] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1085.288525] env[61991]: value = "task-1130408" [ 1085.288525] env[61991]: _type = "Task" [ 1085.288525] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.298415] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130408, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.521465] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.521744] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.521972] env[61991]: DEBUG nova.objects.instance [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'resources' on Instance uuid a0901469-3a2a-4d91-90fc-1909c0573caf {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.749058] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 7cbf26bd-51ab-41f7-976a-4ad9c3c1a997] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.798423] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130408, 'name': ReconfigVM_Task, 'duration_secs': 0.171113} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.798686] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1085.799524] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e60174e-0e2e-4ac6-bc93-93adff3a7572 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.820859] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.821140] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d896ba0f-0a7f-4290-8814-86be4fa8447a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.838316] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1085.838316] env[61991]: value = "task-1130409" [ 1085.838316] env[61991]: _type = "Task" [ 1085.838316] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.845685] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130409, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.121085] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec026dc0-cff1-4b5b-9585-6a809af30517 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.128707] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae18d6ee-7884-4d83-8240-38b2c0ac9c25 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.157250] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9958f4-9b67-4120-a13a-1e9e22280ac6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.164050] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f61887c-8f34-45e2-abea-6f126c1d78c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.177904] env[61991]: DEBUG nova.compute.provider_tree [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.251526] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: d2a0e5d1-5557-41b6-936d-dc86d1346c61] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.348281] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130409, 'name': ReconfigVM_Task, 'duration_secs': 0.337173} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.348641] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.349096] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 50 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1086.681511] env[61991]: DEBUG nova.scheduler.client.report [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1086.754801] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 7df78da9-bf65-4621-b50d-43f1d721c2f1] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.857657] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04661d22-5f38-43f4-a7e5-6cb489ae25b1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.875996] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440cca9a-14ae-4646-b8c7-e573de9101c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.892255] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 67 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.186886] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.205288] env[61991]: INFO nova.scheduler.client.report [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance a0901469-3a2a-4d91-90fc-1909c0573caf [ 1087.258213] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 4b3c05a3-c40a-40c8-8501-5c1b8e4ea349] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.434151] env[61991]: DEBUG nova.network.neutron [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Port e0d6f075-39ac-4264-9656-b78feb2b3747 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1087.712233] env[61991]: DEBUG oslo_concurrency.lockutils [None req-87ecf337-f0eb-4e00-884b-c6dfe7f7ad1e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "a0901469-3a2a-4d91-90fc-1909c0573caf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.060s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.761337] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3e06e424-b64e-4fd2-8013-27760200c41d] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.265070] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 700ce4a7-d597-449a-9379-0cfb0c8f82fc] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.455628] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.455856] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.456122] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.715317] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.715522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.768216] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 80d05278-b516-4408-94b0-11bc93500b5c] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.218371] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1089.271485] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 3bcde671-5702-4b8a-8881-88eb7dfd0556] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.489179] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.489387] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.489568] env[61991]: DEBUG nova.network.neutron [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1089.741790] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.742085] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.743519] env[61991]: INFO nova.compute.claims [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1089.773953] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f78ef63b-453e-45d3-959b-4b0c1922b53e] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.245416] env[61991]: DEBUG nova.network.neutron [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.277475] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 1715711a-72c5-4e86-88e0-cecf922cb42f] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.749346] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.780567] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c14d99a3-950d-44d0-b330-3031a1f8a2be] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.851480] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d7b17a-005f-4ed2-b554-7e4a3b4920bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.859059] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd46d1b-63ac-4c6e-b198-b2bce74f6452 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.889733] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48eb356e-cfc2-4f2b-9a94-d14190db5046 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.896634] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939f6cdb-cfee-49ed-8ef4-9ec732dbc1e9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.909132] env[61991]: DEBUG nova.compute.provider_tree [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.285637] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9c17f844-1f0e-4f01-aadc-0f1f75a59d06] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.293918] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98008663-6f91-406d-ba03-c2f96d542790 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.312292] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a92b41-16c2-448d-af6c-1d1342ab6f2e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.323468] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 83 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.411973] env[61991]: DEBUG nova.scheduler.client.report [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.789172] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: acef131a-4263-42ad-964e-bb1e0cb21eec] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.820981] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.821304] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.821500] env[61991]: INFO nova.compute.manager [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Shelving [ 1091.829914] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1091.830611] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9de357ac-59d6-4f11-8e73-d968db6bd507 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.839268] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1091.839268] env[61991]: value = "task-1130410" [ 1091.839268] env[61991]: _type = "Task" [ 1091.839268] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.847290] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.917209] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.175s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.917980] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1092.160379] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.160650] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.160865] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.161086] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.161278] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.163516] env[61991]: INFO nova.compute.manager [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Terminating instance [ 1092.165894] env[61991]: DEBUG nova.compute.manager [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1092.166127] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.166385] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-705b1669-ab54-4a8f-92f1-84f39b83e55a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.174374] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1092.174374] env[61991]: value = "task-1130411" [ 1092.174374] env[61991]: _type = "Task" [ 1092.174374] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.183828] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130411, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.292255] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.292464] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Cleaning up deleted instances with incomplete migration {{(pid=61991) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1092.328529] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.328793] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f075af13-1c24-4c6e-8b96-f41e6ed5406c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.337131] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1092.337131] env[61991]: value = "task-1130412" [ 1092.337131] env[61991]: _type = "Task" [ 1092.337131] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.348046] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130412, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.350807] env[61991]: DEBUG oslo_vmware.api [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130410, 'name': PowerOnVM_Task, 'duration_secs': 0.364119} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.351122] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1092.351738] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1152014c-03f8-4354-a2a2-aa2ec1f348d8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance 'b26f02c6-a8ea-402d-922d-7879ccb28b92' progress to 100 {{(pid=61991) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1092.424673] env[61991]: DEBUG nova.compute.utils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.425776] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1092.426349] env[61991]: DEBUG nova.network.neutron [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1092.466090] env[61991]: DEBUG nova.policy [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1092.683643] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130411, 'name': PowerOffVM_Task, 'duration_secs': 0.251097} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.683958] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1092.684185] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1092.684383] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246962', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'name': 'volume-f529fecd-03fc-4258-a8dc-f770503a03ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'c43df224-7c35-4b82-ba13-50f3a6f93f2e', 'attached_at': '2024-10-04T09:26:22.000000', 'detached_at': '', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'serial': 'f529fecd-03fc-4258-a8dc-f770503a03ee'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1092.685164] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6cbdd6-bcfe-4a85-95b7-110613db4dfa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.702959] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfb24df-aacb-49cd-aef2-6a69b2def61c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.709081] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e488373-8a60-446f-9979-328eaa760937 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.725801] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd4eeb7-0c9c-4a88-a1dc-6dc78ee53402 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.740285] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] The volume has not been displaced from its original location: [datastore2] volume-f529fecd-03fc-4258-a8dc-f770503a03ee/volume-f529fecd-03fc-4258-a8dc-f770503a03ee.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1092.745389] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.745667] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ad750a3-fb8f-4b7b-b014-070138dc585a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.763295] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1092.763295] env[61991]: value = "task-1130413" [ 1092.763295] env[61991]: _type = "Task" [ 1092.763295] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.771525] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130413, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.772391] env[61991]: DEBUG nova.network.neutron [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Successfully created port: 4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1092.795485] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.848372] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130412, 'name': PowerOffVM_Task, 'duration_secs': 0.143971} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.848745] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1092.849588] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dea7467-d9d6-4494-85e1-bcf0cc9e3be5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.871623] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d01056e-5084-4692-9e46-9ab35578fd31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.929506] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1093.273273] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130413, 'name': ReconfigVM_Task, 'duration_secs': 0.179548} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.273589] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1093.278435] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0243cc3-2ddc-4832-9758-6b6fad36ceac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.293220] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1093.293220] env[61991]: value = "task-1130414" [ 1093.293220] env[61991]: _type = "Task" [ 1093.293220] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.300954] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130414, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.383048] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Creating Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1093.383219] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bd027496-14b6-40ce-80d0-1ba545ea6732 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.390695] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1093.390695] env[61991]: value = "task-1130415" [ 1093.390695] env[61991]: _type = "Task" [ 1093.390695] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.398454] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130415, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.804396] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130414, 'name': ReconfigVM_Task, 'duration_secs': 0.163815} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.804396] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246962', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'name': 'volume-f529fecd-03fc-4258-a8dc-f770503a03ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'c43df224-7c35-4b82-ba13-50f3a6f93f2e', 'attached_at': '2024-10-04T09:26:22.000000', 'detached_at': '', 'volume_id': 'f529fecd-03fc-4258-a8dc-f770503a03ee', 'serial': 'f529fecd-03fc-4258-a8dc-f770503a03ee'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1093.804396] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.804396] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04cb097-9913-429b-89b6-b58150167fe7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.810417] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.810979] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d20d7e5a-597f-48bc-863e-ece082cf0281 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.869689] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.869974] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.870213] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore2] c43df224-7c35-4b82-ba13-50f3a6f93f2e {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.870497] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5be404ba-117b-4cd6-b703-b637022b7049 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.883711] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1093.883711] env[61991]: value = "task-1130417" [ 1093.883711] env[61991]: _type = "Task" [ 1093.883711] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.894264] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130417, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.901670] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130415, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.939544] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1093.967304] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1093.967591] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1093.967755] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1093.967947] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1093.968125] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1093.968326] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1093.968563] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1093.968726] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1093.968897] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1093.969078] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1093.969258] env[61991]: DEBUG nova.virt.hardware [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.970465] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89e6fbf-7ed7-4640-8d49-b2cd33fc83ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.977881] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16566134-be13-4a9a-a6eb-5e86592e1096 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.157208] env[61991]: DEBUG nova.compute.manager [req-ef978543-b79f-425f-b700-f11223234d61 req-80ff8675-d66e-4aaf-99d2-96243831e2ee service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Received event network-vif-plugged-4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.157440] env[61991]: DEBUG oslo_concurrency.lockutils [req-ef978543-b79f-425f-b700-f11223234d61 req-80ff8675-d66e-4aaf-99d2-96243831e2ee service nova] Acquiring lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.157654] env[61991]: DEBUG oslo_concurrency.lockutils [req-ef978543-b79f-425f-b700-f11223234d61 req-80ff8675-d66e-4aaf-99d2-96243831e2ee service nova] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.157826] env[61991]: DEBUG oslo_concurrency.lockutils [req-ef978543-b79f-425f-b700-f11223234d61 req-80ff8675-d66e-4aaf-99d2-96243831e2ee service nova] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.157996] env[61991]: DEBUG nova.compute.manager [req-ef978543-b79f-425f-b700-f11223234d61 req-80ff8675-d66e-4aaf-99d2-96243831e2ee service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] No waiting events found dispatching network-vif-plugged-4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1094.158219] env[61991]: WARNING nova.compute.manager [req-ef978543-b79f-425f-b700-f11223234d61 req-80ff8675-d66e-4aaf-99d2-96243831e2ee service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Received unexpected event network-vif-plugged-4ae9140c-c60a-421c-aa42-9c4694eb1fe5 for instance with vm_state building and task_state spawning. [ 1094.277274] env[61991]: DEBUG nova.network.neutron [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Port e0d6f075-39ac-4264-9656-b78feb2b3747 binding to destination host cpu-1 is already ACTIVE {{(pid=61991) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1094.277567] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.277724] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.277889] env[61991]: DEBUG nova.network.neutron [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.295607] env[61991]: DEBUG nova.network.neutron [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Successfully updated port: 4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1094.395340] env[61991]: DEBUG oslo_vmware.api [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130417, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093829} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.398858] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.398999] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1094.399217] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.399444] env[61991]: INFO nova.compute.manager [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1094.399755] env[61991]: DEBUG oslo.service.loopingcall [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.399971] env[61991]: DEBUG nova.compute.manager [-] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.400109] env[61991]: DEBUG nova.network.neutron [-] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.406795] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130415, 'name': CreateSnapshot_Task, 'duration_secs': 0.595724} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.407048] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Created Snapshot of the VM instance {{(pid=61991) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1094.407859] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db881b8f-b653-47b3-8677-9966f0550479 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.797869] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-891eb9f3-d68a-4d10-996b-f2a780a4d3f9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.798384] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-891eb9f3-d68a-4d10-996b-f2a780a4d3f9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.798384] env[61991]: DEBUG nova.network.neutron [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.925608] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Creating linked-clone VM from snapshot {{(pid=61991) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1094.925962] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-14585bac-b801-44c0-a63b-6732a4e737e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.934927] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1094.934927] env[61991]: value = "task-1130418" [ 1094.934927] env[61991]: _type = "Task" [ 1094.934927] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.943465] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130418, 'name': CloneVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.278864] env[61991]: DEBUG nova.network.neutron [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.329745] env[61991]: DEBUG nova.network.neutron [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1095.368537] env[61991]: DEBUG nova.network.neutron [-] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.449484] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130418, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.475556] env[61991]: DEBUG nova.network.neutron [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Updating instance_info_cache with network_info: [{"id": "4ae9140c-c60a-421c-aa42-9c4694eb1fe5", "address": "fa:16:3e:46:50:69", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ae9140c-c6", "ovs_interfaceid": "4ae9140c-c60a-421c-aa42-9c4694eb1fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.781355] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.871151] env[61991]: INFO nova.compute.manager [-] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Took 1.47 seconds to deallocate network for instance. [ 1095.945632] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130418, 'name': CloneVM_Task} progress is 94%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.978455] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-891eb9f3-d68a-4d10-996b-f2a780a4d3f9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.978787] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Instance network_info: |[{"id": "4ae9140c-c60a-421c-aa42-9c4694eb1fe5", "address": "fa:16:3e:46:50:69", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ae9140c-c6", "ovs_interfaceid": "4ae9140c-c60a-421c-aa42-9c4694eb1fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1095.979477] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:50:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ae9140c-c60a-421c-aa42-9c4694eb1fe5', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.987105] env[61991]: DEBUG oslo.service.loopingcall [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.987105] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1095.987253] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6146aa28-f30c-4305-b97b-dab95d5ae97a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.007957] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.007957] env[61991]: value = "task-1130419" [ 1096.007957] env[61991]: _type = "Task" [ 1096.007957] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.015517] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130419, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.187785] env[61991]: DEBUG nova.compute.manager [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Received event network-changed-4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1096.188050] env[61991]: DEBUG nova.compute.manager [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Refreshing instance network info cache due to event network-changed-4ae9140c-c60a-421c-aa42-9c4694eb1fe5. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1096.188330] env[61991]: DEBUG oslo_concurrency.lockutils [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] Acquiring lock "refresh_cache-891eb9f3-d68a-4d10-996b-f2a780a4d3f9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.188529] env[61991]: DEBUG oslo_concurrency.lockutils [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] Acquired lock "refresh_cache-891eb9f3-d68a-4d10-996b-f2a780a4d3f9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.188737] env[61991]: DEBUG nova.network.neutron [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Refreshing network info cache for port 4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1096.284466] env[61991]: DEBUG nova.compute.manager [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61991) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1096.284742] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.285036] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.414814] env[61991]: INFO nova.compute.manager [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Took 0.54 seconds to detach 1 volumes for instance. [ 1096.417528] env[61991]: DEBUG nova.compute.manager [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Deleting volume: f529fecd-03fc-4258-a8dc-f770503a03ee {{(pid=61991) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1096.446054] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130418, 'name': CloneVM_Task} progress is 95%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.518274] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130419, 'name': CreateVM_Task, 'duration_secs': 0.426186} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.518524] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1096.519165] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.519349] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.519712] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1096.519966] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8b4c92c-3356-405a-93c8-e3048a7077e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.524468] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1096.524468] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5299895b-2655-290c-63db-948e48a21fcf" [ 1096.524468] env[61991]: _type = "Task" [ 1096.524468] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.532042] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5299895b-2655-290c-63db-948e48a21fcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.787670] env[61991]: DEBUG nova.objects.instance [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'migration_context' on Instance uuid b26f02c6-a8ea-402d-922d-7879ccb28b92 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.883853] env[61991]: DEBUG nova.network.neutron [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Updated VIF entry in instance network info cache for port 4ae9140c-c60a-421c-aa42-9c4694eb1fe5. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1096.884245] env[61991]: DEBUG nova.network.neutron [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Updating instance_info_cache with network_info: [{"id": "4ae9140c-c60a-421c-aa42-9c4694eb1fe5", "address": "fa:16:3e:46:50:69", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ae9140c-c6", "ovs_interfaceid": "4ae9140c-c60a-421c-aa42-9c4694eb1fe5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.948380] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130418, 'name': CloneVM_Task, 'duration_secs': 1.595696} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.948775] env[61991]: INFO nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Created linked-clone VM from snapshot [ 1096.949407] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a16c07-fb78-48e8-bac7-7a8fc1e8ec9c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.956272] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Uploading image cf4e271b-b1ee-4b85-86b6-9bdddd82b5db {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1096.959131] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.980167] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1096.980167] env[61991]: value = "vm-246974" [ 1096.980167] env[61991]: _type = "VirtualMachine" [ 1096.980167] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1096.980437] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-38973eb0-dfe0-40fa-858d-aebb45472a70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.986413] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease: (returnval){ [ 1096.986413] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df130f-1554-8ead-6c81-9edbe87ac83f" [ 1096.986413] env[61991]: _type = "HttpNfcLease" [ 1096.986413] env[61991]: } obtained for exporting VM: (result){ [ 1096.986413] env[61991]: value = "vm-246974" [ 1096.986413] env[61991]: _type = "VirtualMachine" [ 1096.986413] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1096.986636] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the lease: (returnval){ [ 1096.986636] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df130f-1554-8ead-6c81-9edbe87ac83f" [ 1096.986636] env[61991]: _type = "HttpNfcLease" [ 1096.986636] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1096.992243] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.992243] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df130f-1554-8ead-6c81-9edbe87ac83f" [ 1096.992243] env[61991]: _type = "HttpNfcLease" [ 1096.992243] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1097.033628] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5299895b-2655-290c-63db-948e48a21fcf, 'name': SearchDatastore_Task, 'duration_secs': 0.008011} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.033932] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.034196] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.034434] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.034582] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.034758] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.034996] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f541674f-416b-4a35-84b3-0cdd36dc72b4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.042479] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.042663] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1097.043399] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f77af5f-0393-4181-ad25-508548a3f7bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.048194] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1097.048194] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b5949-8a46-6c78-3a0e-8680122608dd" [ 1097.048194] env[61991]: _type = "Task" [ 1097.048194] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.055511] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b5949-8a46-6c78-3a0e-8680122608dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.386747] env[61991]: DEBUG oslo_concurrency.lockutils [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] Releasing lock "refresh_cache-891eb9f3-d68a-4d10-996b-f2a780a4d3f9" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.387014] env[61991]: DEBUG nova.compute.manager [req-3ae65e29-e04d-4f26-bd5f-e2c8c301c987 req-9102588f-85cd-458a-878c-1613fa27ebc7 service nova] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Received event network-vif-deleted-7277dc2e-f381-41e1-95de-31eda0f6cb27 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.387933] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68477dcd-2362-45db-a2a7-5485628b6bed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.394974] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287e0a13-fc9f-41e7-8afa-002bf19e2b65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.424374] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142c67a0-0e8c-4343-89cd-fa7a2513a7d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.430916] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7273b3b-4581-4971-987d-0d8de7da5b09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.443365] env[61991]: DEBUG nova.compute.provider_tree [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.493260] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1097.493260] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df130f-1554-8ead-6c81-9edbe87ac83f" [ 1097.493260] env[61991]: _type = "HttpNfcLease" [ 1097.493260] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1097.493498] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1097.493498] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52df130f-1554-8ead-6c81-9edbe87ac83f" [ 1097.493498] env[61991]: _type = "HttpNfcLease" [ 1097.493498] env[61991]: }. {{(pid=61991) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1097.494165] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a391c2ae-0a27-4c72-a601-fb733602a080 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.500529] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b031e0-aa83-c1bf-05ae-7ea0e1e9f785/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1097.500705] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b031e0-aa83-c1bf-05ae-7ea0e1e9f785/disk-0.vmdk for reading. {{(pid=61991) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1097.567244] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525b5949-8a46-6c78-3a0e-8680122608dd, 'name': SearchDatastore_Task, 'duration_secs': 0.008982} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.567996] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2223dbb-0882-4fe6-a2d5-773459fc732a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.572522] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1097.572522] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526ef379-9717-5dfa-4ca6-9b053fdcc1d8" [ 1097.572522] env[61991]: _type = "Task" [ 1097.572522] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.579608] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526ef379-9717-5dfa-4ca6-9b053fdcc1d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.606467] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-002eb4df-0d46-41ac-bab1-98b2c4cdf717 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.946717] env[61991]: DEBUG nova.scheduler.client.report [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.083719] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526ef379-9717-5dfa-4ca6-9b053fdcc1d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009425} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.084148] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.084545] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 891eb9f3-d68a-4d10-996b-f2a780a4d3f9/891eb9f3-d68a-4d10-996b-f2a780a4d3f9.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1098.084863] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-376e34f9-a3e1-4783-9721-a2c49815d13e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.092122] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1098.092122] env[61991]: value = "task-1130422" [ 1098.092122] env[61991]: _type = "Task" [ 1098.092122] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.099733] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130422, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.604612] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130422, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.959742] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.675s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.965774] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.007s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.966073] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.986360] env[61991]: INFO nova.scheduler.client.report [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted allocations for instance c43df224-7c35-4b82-ba13-50f3a6f93f2e [ 1099.103519] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130422, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572638} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.103934] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 891eb9f3-d68a-4d10-996b-f2a780a4d3f9/891eb9f3-d68a-4d10-996b-f2a780a4d3f9.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1099.104236] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1099.104882] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1bf6e0b8-f3e1-411a-bc03-c748ef437929 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.111607] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1099.111607] env[61991]: value = "task-1130423" [ 1099.111607] env[61991]: _type = "Task" [ 1099.111607] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.119645] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.494899] env[61991]: DEBUG oslo_concurrency.lockutils [None req-528b7f9e-9820-4e8d-bdc0-3b234620ea95 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "c43df224-7c35-4b82-ba13-50f3a6f93f2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.334s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.622707] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067676} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.622998] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.623927] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f935b97b-c60f-4b13-bf8e-11ea4674ecf3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.647609] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 891eb9f3-d68a-4d10-996b-f2a780a4d3f9/891eb9f3-d68a-4d10-996b-f2a780a4d3f9.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.648356] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88f724db-db5b-41d1-a8ee-e2bcf185aab7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.668612] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1099.668612] env[61991]: value = "task-1130424" [ 1099.668612] env[61991]: _type = "Task" [ 1099.668612] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.676716] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130424, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.007356] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.007820] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.008194] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "100d5fc0-e9d9-4892-9f60-99d657c533ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.008446] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.009031] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.011326] env[61991]: INFO nova.compute.manager [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Terminating instance [ 1100.013281] env[61991]: DEBUG nova.compute.manager [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1100.013458] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1100.014347] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0779e86f-6cfb-4381-a3e8-2fea260b9b69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.022237] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.022577] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e1a8d73-48d4-4994-8d3e-9f02f3b23262 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.029441] env[61991]: DEBUG oslo_vmware.api [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1100.029441] env[61991]: value = "task-1130425" [ 1100.029441] env[61991]: _type = "Task" [ 1100.029441] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.040206] env[61991]: DEBUG oslo_vmware.api [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.178116] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130424, 'name': ReconfigVM_Task, 'duration_secs': 0.427864} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.178622] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 891eb9f3-d68a-4d10-996b-f2a780a4d3f9/891eb9f3-d68a-4d10-996b-f2a780a4d3f9.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.179072] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1dbea762-75b8-4dc6-94d7-2d5aa2012652 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.185224] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1100.185224] env[61991]: value = "task-1130426" [ 1100.185224] env[61991]: _type = "Task" [ 1100.185224] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.196703] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130426, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.292665] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.502923] env[61991]: INFO nova.compute.manager [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Swapping old allocation on dict_keys(['d748992a-e0bf-4ec2-9c17-0e373360e5a3']) held by migration 5d92b469-b2f9-4e3d-9b16-dfe03939edfc for instance [ 1100.525238] env[61991]: DEBUG nova.scheduler.client.report [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Overwriting current allocation {'allocations': {'d748992a-e0bf-4ec2-9c17-0e373360e5a3': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 167}}, 'project_id': '870d2c2c0e554180b190b88bdab5fc2d', 'user_id': 'ceec39441db345a1ac9cc3f9d6f6ee18', 'consumer_generation': 1} on consumer b26f02c6-a8ea-402d-922d-7879ccb28b92 {{(pid=61991) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1100.539878] env[61991]: DEBUG oslo_vmware.api [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130425, 'name': PowerOffVM_Task, 'duration_secs': 0.189303} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.540186] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1100.540358] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1100.540882] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bee45367-cc74-439b-a209-899665bd7608 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.599910] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1100.600209] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1100.600451] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore1] 100d5fc0-e9d9-4892-9f60-99d657c533ed {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1100.600766] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20d1b141-04e6-424e-8957-7ea5b091af11 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.607472] env[61991]: DEBUG oslo_vmware.api [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1100.607472] env[61991]: value = "task-1130428" [ 1100.607472] env[61991]: _type = "Task" [ 1100.607472] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.616958] env[61991]: DEBUG oslo_vmware.api [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.642575] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.642784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.642999] env[61991]: DEBUG nova.network.neutron [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1100.696444] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130426, 'name': Rename_Task, 'duration_secs': 0.168738} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.696706] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.697032] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-79e07eac-2687-4bf5-a047-29b920d0b88b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.702935] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1100.702935] env[61991]: value = "task-1130429" [ 1100.702935] env[61991]: _type = "Task" [ 1100.702935] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.710785] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.799287] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.117881] env[61991]: DEBUG oslo_vmware.api [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.468789} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.118211] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.118694] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1101.118888] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1101.119080] env[61991]: INFO nova.compute.manager [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1101.119341] env[61991]: DEBUG oslo.service.loopingcall [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.119537] env[61991]: DEBUG nova.compute.manager [-] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1101.119633] env[61991]: DEBUG nova.network.neutron [-] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1101.213332] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130429, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.303013] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Getting list of instances from cluster (obj){ [ 1101.303013] env[61991]: value = "domain-c8" [ 1101.303013] env[61991]: _type = "ClusterComputeResource" [ 1101.303013] env[61991]: } {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1101.304100] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2698fecd-50fb-445a-ab69-43bc752e6a1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.318092] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Got total of 5 instances {{(pid=61991) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1101.318291] env[61991]: WARNING nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] While synchronizing instance power states, found 6 instances in the database and 5 instances on the hypervisor. [ 1101.318425] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Triggering sync for uuid cf99c9cc-24c3-4acc-8120-49c4b12a3553 {{(pid=61991) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1101.318615] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Triggering sync for uuid 100d5fc0-e9d9-4892-9f60-99d657c533ed {{(pid=61991) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1101.318770] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Triggering sync for uuid b26f02c6-a8ea-402d-922d-7879ccb28b92 {{(pid=61991) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1101.318920] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Triggering sync for uuid 84a0dee7-c33b-494a-ad38-83da6ab44ce2 {{(pid=61991) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1101.319079] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Triggering sync for uuid 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1101.319234] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Triggering sync for uuid 891eb9f3-d68a-4d10-996b-f2a780a4d3f9 {{(pid=61991) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1101.319580] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.319806] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.320080] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.320301] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.320488] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.320654] env[61991]: INFO nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1101.320817] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.321016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.321207] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.321435] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.321638] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.324719] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33899d9-a56b-4b79-b015-91d834cc48cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.327701] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55271831-61ef-45f1-b851-ef2d19d580eb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.478947] env[61991]: DEBUG nova.compute.manager [req-064caf4b-5e97-44f5-8fe1-f7c01eabdf67 req-f6812601-24a5-4707-b483-9fb461a597fb service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Received event network-vif-deleted-96d11c6e-409f-4baf-983b-c066592aa09b {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.479222] env[61991]: INFO nova.compute.manager [req-064caf4b-5e97-44f5-8fe1-f7c01eabdf67 req-f6812601-24a5-4707-b483-9fb461a597fb service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Neutron deleted interface 96d11c6e-409f-4baf-983b-c066592aa09b; detaching it from the instance and deleting it from the info cache [ 1101.479410] env[61991]: DEBUG nova.network.neutron [req-064caf4b-5e97-44f5-8fe1-f7c01eabdf67 req-f6812601-24a5-4707-b483-9fb461a597fb service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.505899] env[61991]: DEBUG nova.network.neutron [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [{"id": "e0d6f075-39ac-4264-9656-b78feb2b3747", "address": "fa:16:3e:dc:54:f1", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0d6f075-39", "ovs_interfaceid": "e0d6f075-39ac-4264-9656-b78feb2b3747", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.713563] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130429, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.845764] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.846110] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.525s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.952739] env[61991]: DEBUG nova.network.neutron [-] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.982971] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2da2325d-2072-4b6a-b7d6-3832a0d64e74 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.994310] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888d3ec3-b3ce-4c90-b03d-9e6c0e58bd3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.008787] env[61991]: DEBUG oslo_concurrency.lockutils [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-b26f02c6-a8ea-402d-922d-7879ccb28b92" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.009380] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.009668] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3893f243-7fe7-439d-9ca0-1c98214c27ea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.015852] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1102.015852] env[61991]: value = "task-1130430" [ 1102.015852] env[61991]: _type = "Task" [ 1102.015852] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.026222] env[61991]: DEBUG nova.compute.manager [req-064caf4b-5e97-44f5-8fe1-f7c01eabdf67 req-f6812601-24a5-4707-b483-9fb461a597fb service nova] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Detach interface failed, port_id=96d11c6e-409f-4baf-983b-c066592aa09b, reason: Instance 100d5fc0-e9d9-4892-9f60-99d657c533ed could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1102.032357] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.214496] env[61991]: DEBUG oslo_vmware.api [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130429, 'name': PowerOnVM_Task, 'duration_secs': 1.189403} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.214924] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1102.214987] env[61991]: INFO nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Took 8.28 seconds to spawn the instance on the hypervisor. [ 1102.215172] env[61991]: DEBUG nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1102.215977] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44524422-1f58-4f8a-bd80-855ea662bf59 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.455100] env[61991]: INFO nova.compute.manager [-] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Took 1.34 seconds to deallocate network for instance. [ 1102.526228] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130430, 'name': PowerOffVM_Task, 'duration_secs': 0.407011} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.526534] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.527229] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1102.527493] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1102.527712] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1102.527942] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1102.528155] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1102.528365] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1102.528605] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1102.528776] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1102.528946] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1102.529130] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1102.529309] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1102.534237] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c5c3c71-9049-4f8f-8904-ed7a6cf66520 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.549284] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1102.549284] env[61991]: value = "task-1130431" [ 1102.549284] env[61991]: _type = "Task" [ 1102.549284] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.557461] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130431, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.732490] env[61991]: INFO nova.compute.manager [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Took 13.01 seconds to build instance. [ 1102.962272] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.962605] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.962906] env[61991]: DEBUG nova.objects.instance [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'resources' on Instance uuid 100d5fc0-e9d9-4892-9f60-99d657c533ed {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.060243] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130431, 'name': ReconfigVM_Task, 'duration_secs': 0.309794} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.061632] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec3dffc-6947-44bb-9cf2-451bb5cfadf3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.079588] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.079872] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.080054] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.080257] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.080400] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.080553] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.080784] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.080952] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.081141] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.081310] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.081494] env[61991]: DEBUG nova.virt.hardware [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.082320] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24714324-510d-4e30-83fc-917ab92043e0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.087514] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1103.087514] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e04f7e-99d5-5e9a-8e1b-9f778d35b023" [ 1103.087514] env[61991]: _type = "Task" [ 1103.087514] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.096709] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e04f7e-99d5-5e9a-8e1b-9f778d35b023, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.234523] env[61991]: DEBUG oslo_concurrency.lockutils [None req-08a161ee-33cc-4d28-9162-2fe02e0e03cf tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.519s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.234881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.913s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.235022] env[61991]: INFO nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] During sync_power_state the instance has a pending task (spawning). Skip. [ 1103.235360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.592998] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9d6945-8d6e-45f4-bdc5-d14bbef62d39 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.602863] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0ca65b-89f3-45df-adbf-7a5919942dd6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.605927] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e04f7e-99d5-5e9a-8e1b-9f778d35b023, 'name': SearchDatastore_Task, 'duration_secs': 0.010305} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.611230] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1103.611858] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99a774ac-65b4-4d72-81b1-b6aec09dafe5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.648501] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8e5423-6802-4457-968f-d6f10ef84e55 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.656680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.656914] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.658546] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1103.658546] env[61991]: value = "task-1130432" [ 1103.658546] env[61991]: _type = "Task" [ 1103.658546] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.664646] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9367cc77-81a6-4ace-afa7-28894ee87069 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.674888] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.683756] env[61991]: DEBUG nova.compute.provider_tree [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.159026] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1104.171814] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130432, 'name': ReconfigVM_Task, 'duration_secs': 0.321233} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.172176] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1104.172976] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443eabe9-8919-4ff5-9811-7f116e4e3e69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.188921] env[61991]: DEBUG nova.scheduler.client.report [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.200162] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.201223] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae490c16-c3b4-4552-9afe-20097f13d90f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.221822] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1104.221822] env[61991]: value = "task-1130433" [ 1104.221822] env[61991]: _type = "Task" [ 1104.221822] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.232661] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130433, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.682404] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.693521] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.695941] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.014s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.697524] env[61991]: INFO nova.compute.claims [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1104.712665] env[61991]: INFO nova.scheduler.client.report [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted allocations for instance 100d5fc0-e9d9-4892-9f60-99d657c533ed [ 1104.731501] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130433, 'name': ReconfigVM_Task, 'duration_secs': 0.369849} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.731797] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92/b26f02c6-a8ea-402d-922d-7879ccb28b92.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1104.732723] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae8ef7c-6ea0-45b1-b852-131ff90df388 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.753621] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61334d8-3110-4adf-85f8-7ba6dd3c9db7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.773362] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f397efab-c18e-4dff-9590-6d5b658591dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.793558] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236ba6e8-61d3-4a7d-a4a5-a55721c61297 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.800658] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1104.800928] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6088c7a2-6df8-4692-98e3-c8066724a260 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.807805] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1104.807805] env[61991]: value = "task-1130434" [ 1104.807805] env[61991]: _type = "Task" [ 1104.807805] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.815462] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.220955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b85b6078-23f4-44f4-978e-5b98284c21f1 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.213s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.221784] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.902s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.222009] env[61991]: INFO nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] During sync_power_state the instance has a pending task (deleting). Skip. [ 1105.222198] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "100d5fc0-e9d9-4892-9f60-99d657c533ed" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.317629] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130434, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.789626] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35f16f8-1884-477c-b447-352635e80030 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.796877] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c83c0fc-b558-4934-acd9-4e1b998525c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.829958] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c991032a-0ec1-4a61-80c4-55f6779c266d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.837129] env[61991]: DEBUG oslo_vmware.api [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130434, 'name': PowerOnVM_Task, 'duration_secs': 0.840597} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.839239] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1105.844039] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a108703-4a7a-4787-9738-3d4d85e30542 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.857109] env[61991]: DEBUG nova.compute.provider_tree [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.360931] env[61991]: DEBUG nova.scheduler.client.report [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.857297] env[61991]: INFO nova.compute.manager [None req-95dc94a6-716e-49dc-b87a-9c7d4a93af51 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance to original state: 'active' [ 1106.866536] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.867048] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1107.371469] env[61991]: DEBUG nova.compute.utils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.372973] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1107.372973] env[61991]: DEBUG nova.network.neutron [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1107.445352] env[61991]: DEBUG nova.policy [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1107.796286] env[61991]: DEBUG nova.network.neutron [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Successfully created port: 9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1107.875811] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1108.091450] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.091710] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.091927] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.092134] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.092319] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.094288] env[61991]: INFO nova.compute.manager [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Terminating instance [ 1108.096096] env[61991]: DEBUG nova.compute.manager [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1108.096311] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.097152] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd807a9a-a13d-4c9e-ab2d-1e27ccbf1f68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.105782] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.106013] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b01f6d48-4ecc-4018-b029-10226993c0d0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.113835] env[61991]: DEBUG oslo_vmware.api [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1108.113835] env[61991]: value = "task-1130436" [ 1108.113835] env[61991]: _type = "Task" [ 1108.113835] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.121118] env[61991]: DEBUG oslo_vmware.api [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130436, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.624558] env[61991]: DEBUG oslo_vmware.api [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130436, 'name': PowerOffVM_Task, 'duration_secs': 0.408872} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.624834] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1108.625017] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1108.626384] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c65f6e49-3fb9-45c7-9da0-b32d3110bc85 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.704651] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1108.704895] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1108.705071] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleting the datastore file [datastore1] b26f02c6-a8ea-402d-922d-7879ccb28b92 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.705385] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd2b2f7a-f60f-4be9-a3df-579a0de12438 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.708926] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.709160] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.709359] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.709552] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.709742] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.711844] env[61991]: INFO nova.compute.manager [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Terminating instance [ 1108.713590] env[61991]: DEBUG nova.compute.manager [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1108.713820] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.714620] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277f4e3d-8db0-4028-aa6a-184f5b953a09 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.718376] env[61991]: DEBUG oslo_vmware.api [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1108.718376] env[61991]: value = "task-1130438" [ 1108.718376] env[61991]: _type = "Task" [ 1108.718376] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.724155] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.724650] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4f25423-7682-4062-9832-84c5b15b5809 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.728620] env[61991]: DEBUG oslo_vmware.api [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130438, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.733127] env[61991]: DEBUG oslo_vmware.api [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1108.733127] env[61991]: value = "task-1130439" [ 1108.733127] env[61991]: _type = "Task" [ 1108.733127] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.740765] env[61991]: DEBUG oslo_vmware.api [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.877863] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b031e0-aa83-c1bf-05ae-7ea0e1e9f785/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1108.878978] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5c843f-fd0e-4b7e-96b3-0d1c8121b6f6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.886070] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1108.888284] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b031e0-aa83-c1bf-05ae-7ea0e1e9f785/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1108.888562] env[61991]: ERROR oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b031e0-aa83-c1bf-05ae-7ea0e1e9f785/disk-0.vmdk due to incomplete transfer. [ 1108.889052] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-077dedd5-0933-4f12-87d2-41dc6df67469 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.896829] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b031e0-aa83-c1bf-05ae-7ea0e1e9f785/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1108.897055] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Uploaded image cf4e271b-b1ee-4b85-86b6-9bdddd82b5db to the Glance image server {{(pid=61991) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1108.899448] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Destroying the VM {{(pid=61991) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1108.899745] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-48c5a0b4-37da-404d-adbc-37becdd5dbe6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.905688] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1108.905688] env[61991]: value = "task-1130440" [ 1108.905688] env[61991]: _type = "Task" [ 1108.905688] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.911234] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.911473] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.911636] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.911826] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.911974] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.912145] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.912358] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.912526] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.912689] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.912856] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.913040] env[61991]: DEBUG nova.virt.hardware [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.913885] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfad8af-9c6b-492a-b661-96158d718174 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.921821] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130440, 'name': Destroy_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.925159] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38543397-9e03-4fbd-ad07-54581329e7df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.215416] env[61991]: DEBUG nova.compute.manager [req-010f33ff-816a-418b-bdb9-eb6d4ef3aa01 req-a274cf13-bd78-47c4-8e5c-d7b9dd42d81c service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Received event network-vif-plugged-9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.215651] env[61991]: DEBUG oslo_concurrency.lockutils [req-010f33ff-816a-418b-bdb9-eb6d4ef3aa01 req-a274cf13-bd78-47c4-8e5c-d7b9dd42d81c service nova] Acquiring lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.215894] env[61991]: DEBUG oslo_concurrency.lockutils [req-010f33ff-816a-418b-bdb9-eb6d4ef3aa01 req-a274cf13-bd78-47c4-8e5c-d7b9dd42d81c service nova] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.216124] env[61991]: DEBUG oslo_concurrency.lockutils [req-010f33ff-816a-418b-bdb9-eb6d4ef3aa01 req-a274cf13-bd78-47c4-8e5c-d7b9dd42d81c service nova] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.216430] env[61991]: DEBUG nova.compute.manager [req-010f33ff-816a-418b-bdb9-eb6d4ef3aa01 req-a274cf13-bd78-47c4-8e5c-d7b9dd42d81c service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] No waiting events found dispatching network-vif-plugged-9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1109.216744] env[61991]: WARNING nova.compute.manager [req-010f33ff-816a-418b-bdb9-eb6d4ef3aa01 req-a274cf13-bd78-47c4-8e5c-d7b9dd42d81c service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Received unexpected event network-vif-plugged-9d88c5a1-ea75-437c-a22b-473c0bd0a363 for instance with vm_state building and task_state spawning. [ 1109.228877] env[61991]: DEBUG oslo_vmware.api [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130438, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.508034} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.229124] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.229310] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.229527] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.229739] env[61991]: INFO nova.compute.manager [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1109.229975] env[61991]: DEBUG oslo.service.loopingcall [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.230354] env[61991]: DEBUG nova.compute.manager [-] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.230354] env[61991]: DEBUG nova.network.neutron [-] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1109.241397] env[61991]: DEBUG oslo_vmware.api [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130439, 'name': PowerOffVM_Task, 'duration_secs': 0.232374} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.241675] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.241864] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.242425] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8eee07d-f655-464b-b82c-866f97857f92 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.300727] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.300974] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.301181] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleting the datastore file [datastore2] cf99c9cc-24c3-4acc-8120-49c4b12a3553 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.301452] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06e0d39e-acb1-4d45-b351-fd78004e1a12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.308239] env[61991]: DEBUG oslo_vmware.api [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for the task: (returnval){ [ 1109.308239] env[61991]: value = "task-1130442" [ 1109.308239] env[61991]: _type = "Task" [ 1109.308239] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.316506] env[61991]: DEBUG oslo_vmware.api [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130442, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.370922] env[61991]: DEBUG nova.network.neutron [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Successfully updated port: 9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1109.416164] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130440, 'name': Destroy_Task} progress is 33%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.818150] env[61991]: DEBUG oslo_vmware.api [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Task: {'id': task-1130442, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300634} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.818455] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.818659] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.818847] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.819032] env[61991]: INFO nova.compute.manager [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1109.819285] env[61991]: DEBUG oslo.service.loopingcall [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.819478] env[61991]: DEBUG nova.compute.manager [-] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.819591] env[61991]: DEBUG nova.network.neutron [-] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1109.873348] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.873518] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.873632] env[61991]: DEBUG nova.network.neutron [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1109.920870] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130440, 'name': Destroy_Task, 'duration_secs': 0.745035} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.921303] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Destroyed the VM [ 1109.921926] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deleting Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1109.921926] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-19f1e21c-1f78-4931-855e-a535f1738c61 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.930199] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1109.930199] env[61991]: value = "task-1130443" [ 1109.930199] env[61991]: _type = "Task" [ 1109.930199] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.941037] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130443, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.275802] env[61991]: DEBUG nova.network.neutron [-] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.406020] env[61991]: DEBUG nova.network.neutron [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1110.439967] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130443, 'name': RemoveSnapshot_Task, 'duration_secs': 0.354698} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.442329] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deleted Snapshot of the VM instance {{(pid=61991) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1110.442613] env[61991]: DEBUG nova.compute.manager [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1110.443460] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c20d1ac-089d-440a-9fa6-8932cb46f1b8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.541392] env[61991]: DEBUG nova.network.neutron [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Updating instance_info_cache with network_info: [{"id": "9d88c5a1-ea75-437c-a22b-473c0bd0a363", "address": "fa:16:3e:98:88:4b", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d88c5a1-ea", "ovs_interfaceid": "9d88c5a1-ea75-437c-a22b-473c0bd0a363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.543988] env[61991]: DEBUG nova.network.neutron [-] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.779185] env[61991]: INFO nova.compute.manager [-] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Took 1.55 seconds to deallocate network for instance. [ 1110.958990] env[61991]: INFO nova.compute.manager [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Shelve offloading [ 1110.960993] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.961258] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cc3857f-479d-439c-ac4e-2719392d8d31 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.967376] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1110.967376] env[61991]: value = "task-1130444" [ 1110.967376] env[61991]: _type = "Task" [ 1110.967376] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.977538] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1110.977733] env[61991]: DEBUG nova.compute.manager [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1110.978496] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b04ea2-2372-4aab-b5be-f21d38fb4f20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.984040] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.984220] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.984395] env[61991]: DEBUG nova.network.neutron [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1111.044289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.044604] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Instance network_info: |[{"id": "9d88c5a1-ea75-437c-a22b-473c0bd0a363", "address": "fa:16:3e:98:88:4b", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d88c5a1-ea", "ovs_interfaceid": "9d88c5a1-ea75-437c-a22b-473c0bd0a363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1111.045036] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:88:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d88c5a1-ea75-437c-a22b-473c0bd0a363', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1111.052868] env[61991]: DEBUG oslo.service.loopingcall [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.053215] env[61991]: INFO nova.compute.manager [-] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Took 1.23 seconds to deallocate network for instance. [ 1111.053423] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1111.055006] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1ce0be9-d3cc-4fd2-9000-084802bb920c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.076766] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1111.076766] env[61991]: value = "task-1130445" [ 1111.076766] env[61991]: _type = "Task" [ 1111.076766] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.086934] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130445, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.245106] env[61991]: DEBUG nova.compute.manager [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Received event network-changed-9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.245479] env[61991]: DEBUG nova.compute.manager [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Refreshing instance network info cache due to event network-changed-9d88c5a1-ea75-437c-a22b-473c0bd0a363. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1111.245541] env[61991]: DEBUG oslo_concurrency.lockutils [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] Acquiring lock "refresh_cache-2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.245667] env[61991]: DEBUG oslo_concurrency.lockutils [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] Acquired lock "refresh_cache-2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.245833] env[61991]: DEBUG nova.network.neutron [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Refreshing network info cache for port 9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1111.285292] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.285725] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.286072] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.308403] env[61991]: INFO nova.scheduler.client.report [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted allocations for instance b26f02c6-a8ea-402d-922d-7879ccb28b92 [ 1111.492409] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1111.572995] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.573474] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.573873] env[61991]: DEBUG nova.objects.instance [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lazy-loading 'resources' on Instance uuid cf99c9cc-24c3-4acc-8120-49c4b12a3553 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.587193] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130445, 'name': CreateVM_Task, 'duration_secs': 0.318916} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.587467] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1111.588184] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.588550] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.588800] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1111.589334] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82137fa1-d2f1-41fa-b8f1-95231e399e08 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.594268] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1111.594268] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f074ea-a89c-482f-a00b-69310312e7f3" [ 1111.594268] env[61991]: _type = "Task" [ 1111.594268] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.602032] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f074ea-a89c-482f-a00b-69310312e7f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.719505] env[61991]: DEBUG nova.network.neutron [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.816262] env[61991]: DEBUG oslo_concurrency.lockutils [None req-719466ec-ec65-48ec-a429-7a0409c07c94 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "b26f02c6-a8ea-402d-922d-7879ccb28b92" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.724s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.969289] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1111.973135] env[61991]: DEBUG nova.network.neutron [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Updated VIF entry in instance network info cache for port 9d88c5a1-ea75-437c-a22b-473c0bd0a363. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1111.973483] env[61991]: DEBUG nova.network.neutron [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Updating instance_info_cache with network_info: [{"id": "9d88c5a1-ea75-437c-a22b-473c0bd0a363", "address": "fa:16:3e:98:88:4b", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d88c5a1-ea", "ovs_interfaceid": "9d88c5a1-ea75-437c-a22b-473c0bd0a363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.106958] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f074ea-a89c-482f-a00b-69310312e7f3, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.107185] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.107750] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1112.107998] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.108226] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.108416] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1112.108706] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6ea6ef9-e1cb-4460-9a63-6dd518cd3e3f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.121095] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1112.121289] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1112.124053] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc187cb6-e2ab-411a-8f6c-65a73db19e40 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.130030] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1112.130030] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db49ba-84df-d84c-216c-75cd151c4246" [ 1112.130030] env[61991]: _type = "Task" [ 1112.130030] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.136896] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db49ba-84df-d84c-216c-75cd151c4246, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.158419] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff459fe4-798b-4383-b75f-8ef206162f7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.165123] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7a80ff-4c86-49c5-b91f-ccde15acdccc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.195269] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab786580-293a-42db-b1cc-c41a8bd9a809 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.202126] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322f523a-4a61-48b4-a4d7-ecda1fc6d3a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.215050] env[61991]: DEBUG nova.compute.provider_tree [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.221955] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.476110] env[61991]: DEBUG oslo_concurrency.lockutils [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] Releasing lock "refresh_cache-2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.476414] env[61991]: DEBUG nova.compute.manager [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Received event network-vif-deleted-e0d6f075-39ac-4264-9656-b78feb2b3747 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.476612] env[61991]: DEBUG nova.compute.manager [req-78d5b30d-a0d0-4e45-bb80-33726b941320 req-649e2025-d8c3-4e4d-9e7b-bd5a3852e571 service nova] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Received event network-vif-deleted-8fdca45f-c9aa-4a3c-b40f-c1a01fc6be15 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.543408] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1112.544456] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dfff8a-d8bd-4dbb-b963-43e29f20d067 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.552360] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1112.552628] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3119bdb1-f6d6-46b6-8b9b-010aeb9e815c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.612580] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1112.612814] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1112.612990] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleting the datastore file [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1112.613290] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2ba0400-d348-4234-bc6f-de835ae81812 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.619475] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1112.619475] env[61991]: value = "task-1130447" [ 1112.619475] env[61991]: _type = "Task" [ 1112.619475] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.627326] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.637088] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db49ba-84df-d84c-216c-75cd151c4246, 'name': SearchDatastore_Task, 'duration_secs': 0.013483} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.637867] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8615bc8f-0cae-41c7-8a84-d215fd0574d2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.642854] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1112.642854] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fe7b3c-fb7c-acf7-5f34-6ef2b2df3996" [ 1112.642854] env[61991]: _type = "Task" [ 1112.642854] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.651192] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fe7b3c-fb7c-acf7-5f34-6ef2b2df3996, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.718535] env[61991]: DEBUG nova.scheduler.client.report [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.970514] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1113.033900] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "f42c1c55-3635-460e-a106-e63e57426fd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.034192] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.128750] env[61991]: DEBUG oslo_vmware.api [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133237} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.128998] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.129198] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1113.129377] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1113.151533] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52fe7b3c-fb7c-acf7-5f34-6ef2b2df3996, 'name': SearchDatastore_Task, 'duration_secs': 0.008505} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.152396] env[61991]: INFO nova.scheduler.client.report [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted allocations for instance 22f59842-073b-4bca-bf4f-a83552a90582 [ 1113.154800] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.155612] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d/2d4a14e9-0ea7-42bf-b35f-ce8d374c489d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1113.155612] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a5dbe3f-5ed6-433f-bd62-3a38676b69e7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.161722] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1113.161722] env[61991]: value = "task-1130448" [ 1113.161722] env[61991]: _type = "Task" [ 1113.161722] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.169121] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130448, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.223466] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.240730] env[61991]: INFO nova.scheduler.client.report [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Deleted allocations for instance cf99c9cc-24c3-4acc-8120-49c4b12a3553 [ 1113.273935] env[61991]: DEBUG nova.compute.manager [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-vif-unplugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1113.274210] env[61991]: DEBUG oslo_concurrency.lockutils [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.274463] env[61991]: DEBUG oslo_concurrency.lockutils [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.274690] env[61991]: DEBUG oslo_concurrency.lockutils [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.274904] env[61991]: DEBUG nova.compute.manager [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] No waiting events found dispatching network-vif-unplugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1113.275143] env[61991]: WARNING nova.compute.manager [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received unexpected event network-vif-unplugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 for instance with vm_state shelved_offloaded and task_state None. [ 1113.275352] env[61991]: DEBUG nova.compute.manager [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1113.275549] env[61991]: DEBUG nova.compute.manager [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing instance network info cache due to event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1113.275775] env[61991]: DEBUG oslo_concurrency.lockutils [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.275946] env[61991]: DEBUG oslo_concurrency.lockutils [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.276632] env[61991]: DEBUG nova.network.neutron [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1113.536480] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1113.657463] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.657759] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.657990] env[61991]: DEBUG nova.objects.instance [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'resources' on Instance uuid 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.672184] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130448, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438019} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.672422] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d/2d4a14e9-0ea7-42bf-b35f-ce8d374c489d.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1113.672631] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1113.672862] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cfb4fe3-285e-4b6a-a664-9fbbcc0b9a48 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.679640] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1113.679640] env[61991]: value = "task-1130449" [ 1113.679640] env[61991]: _type = "Task" [ 1113.679640] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.686632] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130449, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.747635] env[61991]: DEBUG oslo_concurrency.lockutils [None req-162c2546-91af-4820-9975-c29d6c4bb243 tempest-ServerActionsTestOtherA-236281420 tempest-ServerActionsTestOtherA-236281420-project-member] Lock "cf99c9cc-24c3-4acc-8120-49c4b12a3553" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.038s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.969899] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1113.970090] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1113.970281] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.057927] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.078011] env[61991]: DEBUG nova.network.neutron [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updated VIF entry in instance network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1114.078380] env[61991]: DEBUG nova.network.neutron [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": null, "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.160325] env[61991]: DEBUG nova.objects.instance [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'numa_topology' on Instance uuid 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.188726] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130449, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070086} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.189015] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1114.189826] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89555bcc-49d2-4924-8740-ee38a84f9120 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.211886] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d/2d4a14e9-0ea7-42bf-b35f-ce8d374c489d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.212169] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c67ed70-f18e-4694-a674-236e1d0f148c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.231232] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1114.231232] env[61991]: value = "task-1130450" [ 1114.231232] env[61991]: _type = "Task" [ 1114.231232] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.239095] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130450, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.474341] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.581429] env[61991]: DEBUG oslo_concurrency.lockutils [req-33a245f7-48a6-4922-9d52-21ac1af18377 req-39aa2c08-ce90-44e3-af4b-b76546fdb1bd service nova] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.663328] env[61991]: DEBUG nova.objects.base [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Object Instance<22f59842-073b-4bca-bf4f-a83552a90582> lazy-loaded attributes: resources,numa_topology {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1114.737461] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd15c0c7-4436-4889-8d70-4e2e3169f940 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.744684] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130450, 'name': ReconfigVM_Task, 'duration_secs': 0.297015} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.745343] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d/2d4a14e9-0ea7-42bf-b35f-ce8d374c489d.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.745959] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38ef674d-8f1c-4cec-bfa9-eed0f83a43e4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.750034] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff252989-8403-46ee-8eec-8ffda099d1ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.753767] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1114.753767] env[61991]: value = "task-1130451" [ 1114.753767] env[61991]: _type = "Task" [ 1114.753767] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.782007] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b778b4bc-a554-4bb8-9520-ea7f8b5ba9ed {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.787691] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130451, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.792504] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1442a9af-ba24-46e8-830d-eb23a7f64e84 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.805395] env[61991]: DEBUG nova.compute.provider_tree [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1115.265259] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130451, 'name': Rename_Task, 'duration_secs': 0.142202} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.265612] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1115.265870] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2ebf40d-ffba-4806-a5ae-db3b1c79f917 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.271585] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1115.271585] env[61991]: value = "task-1130452" [ 1115.271585] env[61991]: _type = "Task" [ 1115.271585] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.279956] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130452, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.308229] env[61991]: DEBUG nova.scheduler.client.report [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.432101] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.781008] env[61991]: DEBUG oslo_vmware.api [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130452, 'name': PowerOnVM_Task, 'duration_secs': 0.466649} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.781294] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1115.781501] env[61991]: INFO nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Took 6.90 seconds to spawn the instance on the hypervisor. [ 1115.781681] env[61991]: DEBUG nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.782484] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b0b500-0b17-475a-8044-733326f1e100 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.813910] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.156s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.816901] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.759s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.818484] env[61991]: INFO nova.compute.claims [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.299649] env[61991]: INFO nova.compute.manager [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Took 11.63 seconds to build instance. [ 1116.328990] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea136d4b-fa7c-4cf9-8721-768e1b25b86e tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.507s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.330487] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "22f59842-073b-4bca-bf4f-a83552a90582" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.009s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.330487] env[61991]: INFO nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] During sync_power_state the instance has a pending task (shelving_image_uploading). Skip. [ 1116.330622] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "22f59842-073b-4bca-bf4f-a83552a90582" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.331289] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.899s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.331423] env[61991]: INFO nova.compute.manager [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Unshelving [ 1116.800734] env[61991]: DEBUG oslo_concurrency.lockutils [None req-2af060e8-fb99-4ca6-8dca-70f07c71a886 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.144s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.913725] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f43a2ca-9a26-451f-937e-e3d0bbd43d16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.921783] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193198cf-4259-4fef-9cc3-5ce43002d944 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.953571] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c6a6b6-bc01-4dea-8d2f-6272ae011ffa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.961410] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f32c022-76f1-4f8d-ac8c-3f5ce31f9cf0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.974835] env[61991]: DEBUG nova.compute.provider_tree [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.273832] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.274135] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.274355] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.274585] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.274770] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.277334] env[61991]: INFO nova.compute.manager [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Terminating instance [ 1117.279178] env[61991]: DEBUG nova.compute.manager [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.279376] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.280208] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1718c71c-e6a0-4a9e-9ec9-05b1d44c07f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.287473] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.287694] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4070186d-c28b-4fcd-a6e1-6af5946727bf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.293655] env[61991]: DEBUG oslo_vmware.api [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1117.293655] env[61991]: value = "task-1130453" [ 1117.293655] env[61991]: _type = "Task" [ 1117.293655] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.300933] env[61991]: DEBUG oslo_vmware.api [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130453, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.361883] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.480022] env[61991]: DEBUG nova.scheduler.client.report [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.803410] env[61991]: DEBUG oslo_vmware.api [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130453, 'name': PowerOffVM_Task, 'duration_secs': 0.1697} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.803840] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1117.804152] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1117.804839] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab08ca14-5253-4a7d-b745-d4635b8af1df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.865070] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1117.865070] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1117.865070] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore1] 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1117.865070] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-add2ed9f-ddbe-4628-abc2-3397607d5244 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.873211] env[61991]: DEBUG oslo_vmware.api [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1117.873211] env[61991]: value = "task-1130455" [ 1117.873211] env[61991]: _type = "Task" [ 1117.873211] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.878675] env[61991]: DEBUG oslo_vmware.api [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.986162] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.986162] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1117.988522] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.514s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.988907] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.989229] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1117.989661] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.628s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.990059] env[61991]: DEBUG nova.objects.instance [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'pci_requests' on Instance uuid 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.991931] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26a943f2-a1bf-405a-9718-1685dd5c483e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.000162] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e25b960-3f79-45fe-b624-6a3123e2ac9a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.017410] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf526df-7918-4d1f-858b-ae8f2254cdd8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.024650] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa8ec33-907e-4c8c-915a-a0fb991503b9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.057923] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180223MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1118.057923] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.385115] env[61991]: DEBUG oslo_vmware.api [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147602} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.385115] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1118.385115] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1118.385115] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1118.385115] env[61991]: INFO nova.compute.manager [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1118.385115] env[61991]: DEBUG oslo.service.loopingcall [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1118.385115] env[61991]: DEBUG nova.compute.manager [-] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1118.385115] env[61991]: DEBUG nova.network.neutron [-] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1118.493457] env[61991]: DEBUG nova.compute.utils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1118.494738] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1118.498055] env[61991]: DEBUG nova.network.neutron [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1118.502011] env[61991]: DEBUG nova.objects.instance [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'numa_topology' on Instance uuid 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.582601] env[61991]: DEBUG nova.policy [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ceec39441db345a1ac9cc3f9d6f6ee18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '870d2c2c0e554180b190b88bdab5fc2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1118.769684] env[61991]: DEBUG nova.compute.manager [req-6287ad36-dd4e-4e10-9dd2-d5655b8a582a req-922bb097-b7db-4301-b0c4-34cfd4afcbf8 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Received event network-vif-deleted-9d88c5a1-ea75-437c-a22b-473c0bd0a363 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.769909] env[61991]: INFO nova.compute.manager [req-6287ad36-dd4e-4e10-9dd2-d5655b8a582a req-922bb097-b7db-4301-b0c4-34cfd4afcbf8 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Neutron deleted interface 9d88c5a1-ea75-437c-a22b-473c0bd0a363; detaching it from the instance and deleting it from the info cache [ 1118.770104] env[61991]: DEBUG nova.network.neutron [req-6287ad36-dd4e-4e10-9dd2-d5655b8a582a req-922bb097-b7db-4301-b0c4-34cfd4afcbf8 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.001851] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1119.004408] env[61991]: INFO nova.compute.claims [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.022794] env[61991]: DEBUG nova.network.neutron [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Successfully created port: b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1119.178703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.178703] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.210808] env[61991]: DEBUG nova.network.neutron [-] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.273983] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4f5a08c-98cc-4b1f-88e0-a67e49cac442 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.284157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c05fc63-eaf5-4639-a542-a67c7a36102c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.312061] env[61991]: DEBUG nova.compute.manager [req-6287ad36-dd4e-4e10-9dd2-d5655b8a582a req-922bb097-b7db-4301-b0c4-34cfd4afcbf8 service nova] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Detach interface failed, port_id=9d88c5a1-ea75-437c-a22b-473c0bd0a363, reason: Instance 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1119.681299] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1119.713321] env[61991]: INFO nova.compute.manager [-] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Took 1.33 seconds to deallocate network for instance. [ 1120.014824] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1120.043807] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1120.044256] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1120.044538] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.044847] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1120.045106] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.045374] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1120.045703] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1120.045987] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1120.046297] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1120.046585] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1120.046874] env[61991]: DEBUG nova.virt.hardware [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1120.049017] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98311934-808f-4d4c-9836-72f7b0258702 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.057766] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3808fb-1bd2-4435-a264-8a4abbb9a663 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.123910] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f33abaf-6c42-4bf3-822f-2929eef44659 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.131500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae46d70-88d6-4179-9dab-34624b0cf82e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.160767] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7921853e-b90b-409a-8f2e-9a6d8fb383e6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.169043] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8753961e-04ae-4a12-8087-0801818b4d54 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.184812] env[61991]: DEBUG nova.compute.provider_tree [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.198610] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.220620] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.555033] env[61991]: DEBUG nova.network.neutron [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Successfully updated port: b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1120.689981] env[61991]: DEBUG nova.scheduler.client.report [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.800355] env[61991]: DEBUG nova.compute.manager [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Received event network-vif-plugged-b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.800560] env[61991]: DEBUG oslo_concurrency.lockutils [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] Acquiring lock "f42c1c55-3635-460e-a106-e63e57426fd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.800734] env[61991]: DEBUG oslo_concurrency.lockutils [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] Lock "f42c1c55-3635-460e-a106-e63e57426fd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.800867] env[61991]: DEBUG oslo_concurrency.lockutils [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] Lock "f42c1c55-3635-460e-a106-e63e57426fd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.801074] env[61991]: DEBUG nova.compute.manager [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] No waiting events found dispatching network-vif-plugged-b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1120.801226] env[61991]: WARNING nova.compute.manager [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Received unexpected event network-vif-plugged-b5f87d87-c6b0-45c2-a7d6-92438277b844 for instance with vm_state building and task_state spawning. [ 1120.801388] env[61991]: DEBUG nova.compute.manager [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Received event network-changed-b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.801543] env[61991]: DEBUG nova.compute.manager [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Refreshing instance network info cache due to event network-changed-b5f87d87-c6b0-45c2-a7d6-92438277b844. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1120.801730] env[61991]: DEBUG oslo_concurrency.lockutils [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] Acquiring lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.801867] env[61991]: DEBUG oslo_concurrency.lockutils [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] Acquired lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.802036] env[61991]: DEBUG nova.network.neutron [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Refreshing network info cache for port b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1121.058779] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.194736] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.205s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.197298] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.140s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.243282] env[61991]: INFO nova.network.neutron [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1121.365119] env[61991]: DEBUG nova.network.neutron [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1121.484750] env[61991]: DEBUG nova.network.neutron [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.987436] env[61991]: DEBUG oslo_concurrency.lockutils [req-7467beac-8559-4265-8bf8-b16ab02d9a20 req-ac8b4171-409f-45e6-b3fe-8bfbf156d9e5 service nova] Releasing lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.987853] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.988042] env[61991]: DEBUG nova.network.neutron [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1122.224285] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 84a0dee7-c33b-494a-ad38-83da6ab44ce2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1122.224447] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 891eb9f3-d68a-4d10-996b-f2a780a4d3f9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1122.224595] env[61991]: WARNING nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1122.224714] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f42c1c55-3635-460e-a106-e63e57426fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1122.224826] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 22f59842-073b-4bca-bf4f-a83552a90582 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1122.517397] env[61991]: DEBUG nova.network.neutron [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1122.695927] env[61991]: DEBUG nova.network.neutron [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [{"id": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "address": "fa:16:3e:c0:1c:8c", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5f87d87-c6", "ovs_interfaceid": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.727736] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f2c347a3-47ef-48be-8aa6-1b45bc09e7c7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1122.728015] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1122.728169] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1122.822247] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df8c791-af89-43be-bbec-ca87f51db3a3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.830226] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ad1022-c719-4426-8fd3-6250e5924878 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.835014] env[61991]: DEBUG nova.compute.manager [req-c29658f9-b46e-4bd9-ac81-42edbf63dd63 req-4c48484a-937c-4bee-a687-a34d14ef3766 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-vif-plugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.835234] env[61991]: DEBUG oslo_concurrency.lockutils [req-c29658f9-b46e-4bd9-ac81-42edbf63dd63 req-4c48484a-937c-4bee-a687-a34d14ef3766 service nova] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.835442] env[61991]: DEBUG oslo_concurrency.lockutils [req-c29658f9-b46e-4bd9-ac81-42edbf63dd63 req-4c48484a-937c-4bee-a687-a34d14ef3766 service nova] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.835611] env[61991]: DEBUG oslo_concurrency.lockutils [req-c29658f9-b46e-4bd9-ac81-42edbf63dd63 req-4c48484a-937c-4bee-a687-a34d14ef3766 service nova] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.835821] env[61991]: DEBUG nova.compute.manager [req-c29658f9-b46e-4bd9-ac81-42edbf63dd63 req-4c48484a-937c-4bee-a687-a34d14ef3766 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] No waiting events found dispatching network-vif-plugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1122.835937] env[61991]: WARNING nova.compute.manager [req-c29658f9-b46e-4bd9-ac81-42edbf63dd63 req-4c48484a-937c-4bee-a687-a34d14ef3766 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received unexpected event network-vif-plugged-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 for instance with vm_state shelved_offloaded and task_state spawning. [ 1122.863365] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e752857-ab4d-4871-9d14-f624f622bf78 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.870260] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcddfaca-ebd4-46a0-8c79-e2670e956144 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.883103] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.884761] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.884921] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.885108] env[61991]: DEBUG nova.network.neutron [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.198792] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.199100] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Instance network_info: |[{"id": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "address": "fa:16:3e:c0:1c:8c", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5f87d87-c6", "ovs_interfaceid": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1123.199649] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:1c:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3ccbdbb-8b49-4a26-913f-2a448b72280f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5f87d87-c6b0-45c2-a7d6-92438277b844', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.207014] env[61991]: DEBUG oslo.service.loopingcall [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.207272] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1123.207566] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7782eef6-cf56-4b4d-84e8-55fed30d09ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.228010] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.228010] env[61991]: value = "task-1130456" [ 1123.228010] env[61991]: _type = "Task" [ 1123.228010] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.235543] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130456, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.387490] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1123.648060] env[61991]: DEBUG nova.network.neutron [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.737895] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130456, 'name': CreateVM_Task, 'duration_secs': 0.296837} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.738064] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1123.738699] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.738873] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.739219] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1123.739465] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5236f30-7c96-45a4-a68a-2e5c9e190a65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.744721] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1123.744721] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525ada1e-53ca-a3d8-42c3-32e3b61074f4" [ 1123.744721] env[61991]: _type = "Task" [ 1123.744721] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.752209] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525ada1e-53ca-a3d8-42c3-32e3b61074f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.895053] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1123.895053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.697s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.895053] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.696s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.896227] env[61991]: INFO nova.compute.claims [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.151057] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.175904] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7e45120129b9854cf6ab31983d5d2ffc',container_format='bare',created_at=2024-10-04T09:27:01Z,direct_url=,disk_format='vmdk',id=cf4e271b-b1ee-4b85-86b6-9bdddd82b5db,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1295380278-shelved',owner='822d7e3c678e4defa24bb4d8439a62de',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-04T09:27:18Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.176158] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.176317] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.176545] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.176726] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.176875] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.177093] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.177258] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.177513] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.177697] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.177870] env[61991]: DEBUG nova.virt.hardware [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.178703] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e5e73b-63c9-4ce5-a55b-f6a6f3822f43 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.186786] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7779d0e-791b-4765-84a3-831ab2ce5053 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.200128] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:36:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ead20342-9afa-435e-a22b-b4a903457712', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d2cca63-6f27-4e38-a6b1-6e0eda1401e6', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.207512] env[61991]: DEBUG oslo.service.loopingcall [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.207731] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1124.207925] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a52499bc-c7ce-4ffa-bfb5-7ab75edea293 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.226440] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.226440] env[61991]: value = "task-1130457" [ 1124.226440] env[61991]: _type = "Task" [ 1124.226440] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.233556] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130457, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.254452] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]525ada1e-53ca-a3d8-42c3-32e3b61074f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009605} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.254715] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.254937] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1124.255187] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.255347] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.255537] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1124.255765] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8061019-37c4-4983-b14b-49fac2b69ce3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.262851] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1124.263075] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1124.263783] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2492505-cd26-49d9-ae00-20b695f09406 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.268402] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1124.268402] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f133ff-67a0-8fd2-5a9a-e01f71805703" [ 1124.268402] env[61991]: _type = "Task" [ 1124.268402] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.275831] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f133ff-67a0-8fd2-5a9a-e01f71805703, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.735739] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130457, 'name': CreateVM_Task, 'duration_secs': 0.280602} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.735957] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1124.736671] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.736901] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.737316] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1124.737589] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac58ac18-12d4-4306-a5ab-817beedc4e2a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.741630] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1124.741630] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52afd96c-8d59-d852-1228-bf5656b2d5cb" [ 1124.741630] env[61991]: _type = "Task" [ 1124.741630] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.748475] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52afd96c-8d59-d852-1228-bf5656b2d5cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.776247] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f133ff-67a0-8fd2-5a9a-e01f71805703, 'name': SearchDatastore_Task, 'duration_secs': 0.007625} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.776926] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d88d8e66-28f2-408b-93e9-7a677dddbb23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.781533] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1124.781533] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e061a1-5230-19c0-e01c-7cd9721cc3dd" [ 1124.781533] env[61991]: _type = "Task" [ 1124.781533] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.788349] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e061a1-5230-19c0-e01c-7cd9721cc3dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.861440] env[61991]: DEBUG nova.compute.manager [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.861592] env[61991]: DEBUG nova.compute.manager [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing instance network info cache due to event network-changed-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1124.861892] env[61991]: DEBUG oslo_concurrency.lockutils [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] Acquiring lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.862338] env[61991]: DEBUG oslo_concurrency.lockutils [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] Acquired lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.862578] env[61991]: DEBUG nova.network.neutron [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Refreshing network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1124.893719] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.893934] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.894093] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1124.980058] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825a0f8d-f471-4b4f-bdf6-08824e098d8a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.988664] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c81569-39a4-47e4-8c3d-8985abed9c0c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.016777] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c89df2-7b00-4922-ba03-82740e46e4b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.023556] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cab50b1-38cf-4e09-b39f-ee14cd2e4e69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.036364] env[61991]: DEBUG nova.compute.provider_tree [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.251680] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.251956] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Processing image cf4e271b-b1ee-4b85-86b6-9bdddd82b5db {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.252222] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.252376] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.252555] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.252890] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0911f62e-6717-4572-91b0-3ee57502a154 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.260898] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.261091] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1125.261776] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-814ef072-0d7a-46a9-ab2f-544a740e35ac {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.266533] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1125.266533] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d21d83-a966-4311-89b1-1bd9216a7a99" [ 1125.266533] env[61991]: _type = "Task" [ 1125.266533] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.273406] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d21d83-a966-4311-89b1-1bd9216a7a99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.289772] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e061a1-5230-19c0-e01c-7cd9721cc3dd, 'name': SearchDatastore_Task, 'duration_secs': 0.008837} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.290014] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.290273] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f42c1c55-3635-460e-a106-e63e57426fd2/f42c1c55-3635-460e-a106-e63e57426fd2.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1125.290505] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87850fe5-e8fc-4d12-84cb-f3a1c2bad1b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.297294] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1125.297294] env[61991]: value = "task-1130458" [ 1125.297294] env[61991]: _type = "Task" [ 1125.297294] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.304466] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.539018] env[61991]: DEBUG nova.scheduler.client.report [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.616603] env[61991]: DEBUG nova.network.neutron [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updated VIF entry in instance network info cache for port 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1125.617038] env[61991]: DEBUG nova.network.neutron [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [{"id": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "address": "fa:16:3e:22:36:45", "network": {"id": "e8556cdd-0c40-4cc7-af9d-6af105c90ba8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2009814118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "822d7e3c678e4defa24bb4d8439a62de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ead20342-9afa-435e-a22b-b4a903457712", "external-id": "nsx-vlan-transportzone-912", "segmentation_id": 912, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d2cca63-6f", "ovs_interfaceid": "7d2cca63-6f27-4e38-a6b1-6e0eda1401e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.775913] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Preparing fetch location {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1125.776281] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Fetch image to [datastore2] OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1/OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1.vmdk {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1125.776406] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Downloading stream optimized image cf4e271b-b1ee-4b85-86b6-9bdddd82b5db to [datastore2] OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1/OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1.vmdk on the data store datastore2 as vApp {{(pid=61991) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1125.776501] env[61991]: DEBUG nova.virt.vmwareapi.images [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Downloading image file data cf4e271b-b1ee-4b85-86b6-9bdddd82b5db to the ESX as VM named 'OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1' {{(pid=61991) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1125.820014] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464972} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.820014] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f42c1c55-3635-460e-a106-e63e57426fd2/f42c1c55-3635-460e-a106-e63e57426fd2.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1125.820199] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.820660] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e5934f3-adf9-47ce-b611-97d33aefd2d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.826404] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1125.826404] env[61991]: value = "task-1130459" [ 1125.826404] env[61991]: _type = "Task" [ 1125.826404] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.833664] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130459, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.845158] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1125.845158] env[61991]: value = "resgroup-9" [ 1125.845158] env[61991]: _type = "ResourcePool" [ 1125.845158] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1125.845540] env[61991]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-acd2a322-20c8-4eaa-a3bf-5fa55d656d65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.865482] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease: (returnval){ [ 1125.865482] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5224a0db-7610-0794-7a71-7c312266a8ed" [ 1125.865482] env[61991]: _type = "HttpNfcLease" [ 1125.865482] env[61991]: } obtained for vApp import into resource pool (val){ [ 1125.865482] env[61991]: value = "resgroup-9" [ 1125.865482] env[61991]: _type = "ResourcePool" [ 1125.865482] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1125.865783] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the lease: (returnval){ [ 1125.865783] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5224a0db-7610-0794-7a71-7c312266a8ed" [ 1125.865783] env[61991]: _type = "HttpNfcLease" [ 1125.865783] env[61991]: } to be ready. {{(pid=61991) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1125.871816] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1125.871816] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5224a0db-7610-0794-7a71-7c312266a8ed" [ 1125.871816] env[61991]: _type = "HttpNfcLease" [ 1125.871816] env[61991]: } is initializing. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1125.899300] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Didn't find any instances for network info cache update. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1125.899542] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.899751] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.044231] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.149s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.044755] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1126.047598] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.827s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.047805] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.066901] env[61991]: INFO nova.scheduler.client.report [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d [ 1126.119707] env[61991]: DEBUG oslo_concurrency.lockutils [req-a49ab6c8-6a90-4fdf-8e03-c73c7a09725c req-1e4622cb-255c-4298-a9e1-5cb0c78ffc3b service nova] Releasing lock "refresh_cache-22f59842-073b-4bca-bf4f-a83552a90582" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.337419] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130459, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062594} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.337833] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1126.338433] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e03e87-03d5-425b-b7f0-7b1a356c0fa2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.361400] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] f42c1c55-3635-460e-a106-e63e57426fd2/f42c1c55-3635-460e-a106-e63e57426fd2.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.361789] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0ac5502-9768-41d6-b3ac-c99a4533d4da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.383097] env[61991]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1126.383097] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5224a0db-7610-0794-7a71-7c312266a8ed" [ 1126.383097] env[61991]: _type = "HttpNfcLease" [ 1126.383097] env[61991]: } is ready. {{(pid=61991) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1126.384285] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1126.384285] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5224a0db-7610-0794-7a71-7c312266a8ed" [ 1126.384285] env[61991]: _type = "HttpNfcLease" [ 1126.384285] env[61991]: }. {{(pid=61991) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1126.384574] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1126.384574] env[61991]: value = "task-1130461" [ 1126.384574] env[61991]: _type = "Task" [ 1126.384574] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.385235] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53f98cf-25cf-4786-be7b-263f5cb63271 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.394220] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb9cce-1577-20c3-5660-4076f05ad718/disk-0.vmdk from lease info. {{(pid=61991) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1126.394403] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb9cce-1577-20c3-5660-4076f05ad718/disk-0.vmdk. {{(pid=61991) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1126.398524] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130461, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.461584] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-df1b1f9b-4e11-4a89-bf18-de9ffe1577aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.551896] env[61991]: DEBUG nova.compute.utils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1126.553357] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1126.553526] env[61991]: DEBUG nova.network.neutron [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1126.575256] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6eee3679-ccdc-436c-ac7d-19ff26011151 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "2d4a14e9-0ea7-42bf-b35f-ce8d374c489d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.301s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.598982] env[61991]: DEBUG nova.policy [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6262af74d656477fb4462edda64c1be6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f9bec63b7334b9bac22340eca5410f5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1126.898302] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130461, 'name': ReconfigVM_Task, 'duration_secs': 0.287121} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.900452] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Reconfigured VM instance instance-00000071 to attach disk [datastore2] f42c1c55-3635-460e-a106-e63e57426fd2/f42c1c55-3635-460e-a106-e63e57426fd2.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1126.901240] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92cdcbc8-9687-49bc-b29d-ac8ad74aa9ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.907370] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1126.907370] env[61991]: value = "task-1130462" [ 1126.907370] env[61991]: _type = "Task" [ 1126.907370] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.917223] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130462, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.939202] env[61991]: DEBUG nova.network.neutron [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Successfully created port: e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.057694] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1127.118729] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.119044] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.119328] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.119573] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.119758] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.123436] env[61991]: INFO nova.compute.manager [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Terminating instance [ 1127.125720] env[61991]: DEBUG nova.compute.manager [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1127.125941] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1127.126826] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dd5919-3e16-4b6c-9389-45b4ac9ca392 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.136447] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1127.137924] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4471f704-b4a5-4ea6-ab28-6e5819b05a2e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.144262] env[61991]: DEBUG oslo_vmware.api [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1127.144262] env[61991]: value = "task-1130463" [ 1127.144262] env[61991]: _type = "Task" [ 1127.144262] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.153558] env[61991]: DEBUG oslo_vmware.api [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130463, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.417335] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130462, 'name': Rename_Task, 'duration_secs': 0.151316} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.418586] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1127.418870] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82816828-2d79-4944-935c-214edc2c8f5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.424560] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1127.424560] env[61991]: value = "task-1130464" [ 1127.424560] env[61991]: _type = "Task" [ 1127.424560] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.431948] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130464, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.616705] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Completed reading data from the image iterator. {{(pid=61991) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1127.616959] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb9cce-1577-20c3-5660-4076f05ad718/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1127.618022] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de3ac21-3033-4c4a-860f-0476910e3510 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.624344] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb9cce-1577-20c3-5660-4076f05ad718/disk-0.vmdk is in state: ready. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1127.624518] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb9cce-1577-20c3-5660-4076f05ad718/disk-0.vmdk. {{(pid=61991) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1127.624749] env[61991]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-80d64364-401d-4add-ab08-5d3c7bb45667 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.655753] env[61991]: DEBUG oslo_vmware.api [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130463, 'name': PowerOffVM_Task, 'duration_secs': 0.201309} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.656047] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1127.656228] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1127.656480] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b419b050-56bf-4e10-b1d4-bc31476ab643 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.721697] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1127.721905] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1127.722103] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore2] 891eb9f3-d68a-4d10-996b-f2a780a4d3f9 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.722376] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b345edc-e7cd-4d47-a048-6209f2ae89c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.728804] env[61991]: DEBUG oslo_vmware.api [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1127.728804] env[61991]: value = "task-1130466" [ 1127.728804] env[61991]: _type = "Task" [ 1127.728804] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.736682] env[61991]: DEBUG oslo_vmware.api [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.825677] env[61991]: DEBUG oslo_vmware.rw_handles [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb9cce-1577-20c3-5660-4076f05ad718/disk-0.vmdk. {{(pid=61991) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1127.825677] env[61991]: INFO nova.virt.vmwareapi.images [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Downloaded image file data cf4e271b-b1ee-4b85-86b6-9bdddd82b5db [ 1127.826083] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8136f598-4c53-4aea-84fe-5a9add64f877 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.843318] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0d48ef9-fd41-41f5-bd45-56391f55d38a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.866829] env[61991]: INFO nova.virt.vmwareapi.images [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] The imported VM was unregistered [ 1127.869480] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Caching image {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1127.869729] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Creating directory with path [datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1127.870014] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e16dbd2f-f616-4baf-b871-b628e18d22af {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.883089] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Created directory with path [datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1127.883318] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1/OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1.vmdk to [datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk. {{(pid=61991) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1127.883598] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-931d9a1b-2454-4db9-9a46-3f17c0929340 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.890759] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1127.890759] env[61991]: value = "task-1130468" [ 1127.890759] env[61991]: _type = "Task" [ 1127.890759] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.898917] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.935058] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130464, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.069181] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1128.119447] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.119666] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.119750] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.119996] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.120095] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.120197] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.120407] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.120571] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.120741] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.120900] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.121082] env[61991]: DEBUG nova.virt.hardware [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.122381] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73c9b72-403f-4f9f-96ef-6059bec4528d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.131045] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa7e0b3-8835-4222-9b63-13808dd2302d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.240963] env[61991]: DEBUG oslo_vmware.api [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150449} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.241277] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.241639] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1128.241884] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1128.242082] env[61991]: INFO nova.compute.manager [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1128.242356] env[61991]: DEBUG oslo.service.loopingcall [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.242561] env[61991]: DEBUG nova.compute.manager [-] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1128.242658] env[61991]: DEBUG nova.network.neutron [-] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1128.403220] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.436571] env[61991]: DEBUG oslo_vmware.api [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130464, 'name': PowerOnVM_Task, 'duration_secs': 0.526598} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.436888] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.437102] env[61991]: INFO nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Took 8.42 seconds to spawn the instance on the hypervisor. [ 1128.437317] env[61991]: DEBUG nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.438162] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bddb3e2a-0f3f-4e2c-9623-8eac35fe9767 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.626821] env[61991]: DEBUG nova.compute.manager [req-fa61344c-2203-4099-9282-0c6f5c5e2965 req-4378a861-913a-4e48-b854-f7ba65b2f281 service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Received event network-vif-plugged-e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.627159] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa61344c-2203-4099-9282-0c6f5c5e2965 req-4378a861-913a-4e48-b854-f7ba65b2f281 service nova] Acquiring lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.627566] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa61344c-2203-4099-9282-0c6f5c5e2965 req-4378a861-913a-4e48-b854-f7ba65b2f281 service nova] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.627850] env[61991]: DEBUG oslo_concurrency.lockutils [req-fa61344c-2203-4099-9282-0c6f5c5e2965 req-4378a861-913a-4e48-b854-f7ba65b2f281 service nova] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.628161] env[61991]: DEBUG nova.compute.manager [req-fa61344c-2203-4099-9282-0c6f5c5e2965 req-4378a861-913a-4e48-b854-f7ba65b2f281 service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] No waiting events found dispatching network-vif-plugged-e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1128.628225] env[61991]: WARNING nova.compute.manager [req-fa61344c-2203-4099-9282-0c6f5c5e2965 req-4378a861-913a-4e48-b854-f7ba65b2f281 service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Received unexpected event network-vif-plugged-e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 for instance with vm_state building and task_state spawning. [ 1128.763468] env[61991]: DEBUG nova.network.neutron [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Successfully updated port: e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1128.904559] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.958259] env[61991]: INFO nova.compute.manager [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Took 14.92 seconds to build instance. [ 1129.197469] env[61991]: DEBUG nova.network.neutron [-] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.265652] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "refresh_cache-f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.265865] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquired lock "refresh_cache-f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.266041] env[61991]: DEBUG nova.network.neutron [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1129.404095] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.461229] env[61991]: DEBUG oslo_concurrency.lockutils [None req-802ddde0-bd8c-434b-96a7-ad905a616708 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.427s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.700600] env[61991]: INFO nova.compute.manager [-] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Took 1.46 seconds to deallocate network for instance. [ 1129.810902] env[61991]: DEBUG nova.network.neutron [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1129.905205] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.025038] env[61991]: DEBUG nova.network.neutron [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Updating instance_info_cache with network_info: [{"id": "e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5", "address": "fa:16:3e:c0:9f:fd", "network": {"id": "ab012fbf-ac54-4ecb-ade2-984e895713e1", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1521127739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f9bec63b7334b9bac22340eca5410f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0b63cd0-e6", "ovs_interfaceid": "e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.116483] env[61991]: DEBUG nova.compute.manager [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Received event network-changed-b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.116746] env[61991]: DEBUG nova.compute.manager [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Refreshing instance network info cache due to event network-changed-b5f87d87-c6b0-45c2-a7d6-92438277b844. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1130.117386] env[61991]: DEBUG oslo_concurrency.lockutils [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] Acquiring lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.117642] env[61991]: DEBUG oslo_concurrency.lockutils [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] Acquired lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.117895] env[61991]: DEBUG nova.network.neutron [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Refreshing network info cache for port b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1130.210276] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.210552] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.210778] env[61991]: DEBUG nova.objects.instance [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'resources' on Instance uuid 891eb9f3-d68a-4d10-996b-f2a780a4d3f9 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.403158] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.527613] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Releasing lock "refresh_cache-f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.528024] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Instance network_info: |[{"id": "e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5", "address": "fa:16:3e:c0:9f:fd", "network": {"id": "ab012fbf-ac54-4ecb-ade2-984e895713e1", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1521127739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f9bec63b7334b9bac22340eca5410f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0b63cd0-e6", "ovs_interfaceid": "e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1130.528468] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:9f:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '46e1fc20-2067-4e1a-9812-702772a2c82c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.535815] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Creating folder: Project (9f9bec63b7334b9bac22340eca5410f5). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1130.536120] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9d0558c-4d08-4d8b-85df-87743f7966a5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.548853] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Created folder: Project (9f9bec63b7334b9bac22340eca5410f5) in parent group-v246753. [ 1130.549081] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Creating folder: Instances. Parent ref: group-v246980. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1130.549331] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c433b53-fe6b-44df-af81-6b5579c9b766 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.558833] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Created folder: Instances in parent group-v246980. [ 1130.559045] env[61991]: DEBUG oslo.service.loopingcall [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.559267] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1130.559475] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf1c103d-91e3-424a-96e3-0e587972f163 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.578384] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.578384] env[61991]: value = "task-1130471" [ 1130.578384] env[61991]: _type = "Task" [ 1130.578384] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.588325] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130471, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.665777] env[61991]: DEBUG nova.compute.manager [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Received event network-vif-deleted-4ae9140c-c60a-421c-aa42-9c4694eb1fe5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.665999] env[61991]: DEBUG nova.compute.manager [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Received event network-changed-e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.666214] env[61991]: DEBUG nova.compute.manager [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Refreshing instance network info cache due to event network-changed-e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1130.666437] env[61991]: DEBUG oslo_concurrency.lockutils [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] Acquiring lock "refresh_cache-f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.666583] env[61991]: DEBUG oslo_concurrency.lockutils [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] Acquired lock "refresh_cache-f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.666744] env[61991]: DEBUG nova.network.neutron [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Refreshing network info cache for port e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1130.796222] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002483d1-2442-4314-99cd-72cc7315a362 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.805570] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eeae2a9-7011-4f7a-961c-6ce4ed1dffa8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.837061] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21fb292-89fd-4888-9567-afa3cad2d996 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.844666] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e9ba9c-3f02-4378-88fd-4e6d9692223d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.858662] env[61991]: DEBUG nova.compute.provider_tree [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.873874] env[61991]: DEBUG nova.network.neutron [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updated VIF entry in instance network info cache for port b5f87d87-c6b0-45c2-a7d6-92438277b844. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1130.874268] env[61991]: DEBUG nova.network.neutron [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [{"id": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "address": "fa:16:3e:c0:1c:8c", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5f87d87-c6", "ovs_interfaceid": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.903554] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130468, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.682925} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.903954] env[61991]: INFO nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1/OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1.vmdk to [datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk. [ 1130.904258] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Cleaning up location [datastore2] OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1 {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1130.904527] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_9cafe30b-ce2b-4549-8f3b-10d30798c4c1 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.904851] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff89b62f-14a4-418c-b30d-7e09ad7a6771 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.912247] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1130.912247] env[61991]: value = "task-1130472" [ 1130.912247] env[61991]: _type = "Task" [ 1130.912247] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.920543] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.088153] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130471, 'name': CreateVM_Task, 'duration_secs': 0.377984} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.088353] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1131.089782] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.089983] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.090328] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1131.090583] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0c869e2-fcaa-45d8-8dc7-70fe7f7173d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.094907] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1131.094907] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f187e8-1c54-c142-fc81-a5cbb5422684" [ 1131.094907] env[61991]: _type = "Task" [ 1131.094907] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.102216] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f187e8-1c54-c142-fc81-a5cbb5422684, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.352098] env[61991]: DEBUG nova.network.neutron [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Updated VIF entry in instance network info cache for port e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1131.352585] env[61991]: DEBUG nova.network.neutron [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Updating instance_info_cache with network_info: [{"id": "e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5", "address": "fa:16:3e:c0:9f:fd", "network": {"id": "ab012fbf-ac54-4ecb-ade2-984e895713e1", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1521127739-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f9bec63b7334b9bac22340eca5410f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "46e1fc20-2067-4e1a-9812-702772a2c82c", "external-id": "nsx-vlan-transportzone-210", "segmentation_id": 210, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0b63cd0-e6", "ovs_interfaceid": "e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.361017] env[61991]: DEBUG nova.scheduler.client.report [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.376374] env[61991]: DEBUG oslo_concurrency.lockutils [req-f2c6d90e-ff90-46bf-ac48-ba3d9b5b8e95 req-d7ffd77f-5eda-45b0-aed1-0a800847ed04 service nova] Releasing lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.423746] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042172} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.424076] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.424173] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.424412] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk to [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1131.424653] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ba4a423-dc42-4fc0-870e-715a180b5cb7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.431008] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1131.431008] env[61991]: value = "task-1130473" [ 1131.431008] env[61991]: _type = "Task" [ 1131.431008] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.437762] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130473, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.605537] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52f187e8-1c54-c142-fc81-a5cbb5422684, 'name': SearchDatastore_Task, 'duration_secs': 0.010408} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.605881] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.606178] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.606478] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.606670] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.606894] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.607186] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fac0ade5-e249-4b81-9a28-b09285f6b004 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.615304] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.615495] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.616305] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75b2f7b9-d5f2-49f8-af0c-1922798323a9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.621266] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1131.621266] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205cd30-da4a-4f02-af78-9896b303bea6" [ 1131.621266] env[61991]: _type = "Task" [ 1131.621266] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.628820] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205cd30-da4a-4f02-af78-9896b303bea6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.858228] env[61991]: DEBUG oslo_concurrency.lockutils [req-a2a9a67c-0ead-4341-9a91-29630b3f165d req-cbc055d6-c95a-42d0-a4e5-abe28bee9faa service nova] Releasing lock "refresh_cache-f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.865323] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.655s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.886256] env[61991]: INFO nova.scheduler.client.report [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance 891eb9f3-d68a-4d10-996b-f2a780a4d3f9 [ 1131.943036] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130473, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.133636] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205cd30-da4a-4f02-af78-9896b303bea6, 'name': SearchDatastore_Task, 'duration_secs': 0.008003} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.134476] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f1f8d59-3bf8-4c46-8486-bc088613f516 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.142040] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1132.142040] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e3e231-e367-fa1a-d577-1be376b00ef1" [ 1132.142040] env[61991]: _type = "Task" [ 1132.142040] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.152033] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e3e231-e367-fa1a-d577-1be376b00ef1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.395443] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58df496d-61cf-4658-a16c-059a834460b5 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "891eb9f3-d68a-4d10-996b-f2a780a4d3f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.276s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.443303] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130473, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.653734] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52e3e231-e367-fa1a-d577-1be376b00ef1, 'name': SearchDatastore_Task, 'duration_secs': 0.071276} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.654099] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.654311] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f2c347a3-47ef-48be-8aa6-1b45bc09e7c7/f2c347a3-47ef-48be-8aa6-1b45bc09e7c7.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1132.654588] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de3f7b13-76e9-4bbd-be98-0a3258ee0b0e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.663650] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1132.663650] env[61991]: value = "task-1130474" [ 1132.663650] env[61991]: _type = "Task" [ 1132.663650] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.674908] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.941646] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130473, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.128117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "af7962c8-7700-4ff2-988e-65d2922fd514" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.128358] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "af7962c8-7700-4ff2-988e-65d2922fd514" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.176750] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.443823] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130473, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.631701] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1133.674231] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130474, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.945615] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130473, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.223261} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.945844] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db/cf4e271b-b1ee-4b85-86b6-9bdddd82b5db.vmdk to [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.946676] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3078287-f550-4690-839d-f8c47d65e33c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.968227] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.968494] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db7efe61-adba-4b2c-85db-358c2ecc1406 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.986701] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1133.986701] env[61991]: value = "task-1130475" [ 1133.986701] env[61991]: _type = "Task" [ 1133.986701] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.994399] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130475, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.153840] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.154188] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.155848] env[61991]: INFO nova.compute.claims [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1134.174077] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130474, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.306501} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.174321] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] f2c347a3-47ef-48be-8aa6-1b45bc09e7c7/f2c347a3-47ef-48be-8aa6-1b45bc09e7c7.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1134.174533] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.174779] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01e8f2a7-7920-441b-aa76-501da9b7cb07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.180835] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1134.180835] env[61991]: value = "task-1130476" [ 1134.180835] env[61991]: _type = "Task" [ 1134.180835] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.188957] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130476, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.497531] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.690053] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130476, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.351392} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.690421] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.691175] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ee31ae-c2e5-4e9a-bbce-33a00002a3aa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.713604] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] f2c347a3-47ef-48be-8aa6-1b45bc09e7c7/f2c347a3-47ef-48be-8aa6-1b45bc09e7c7.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.714234] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bad4d12-029c-4f3b-bc59-8aa492ef96ee {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.732637] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1134.732637] env[61991]: value = "task-1130477" [ 1134.732637] env[61991]: _type = "Task" [ 1134.732637] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.740236] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130477, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.998030] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130475, 'name': ReconfigVM_Task, 'duration_secs': 0.712056} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.998335] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582/22f59842-073b-4bca-bf4f-a83552a90582.vmdk or device None with type streamOptimized {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.998994] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a22cb7f7-c82f-4711-bd1c-43afd7dfc29f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.004686] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1135.004686] env[61991]: value = "task-1130478" [ 1135.004686] env[61991]: _type = "Task" [ 1135.004686] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.012077] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130478, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.234022] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0ea30a-1115-460f-a879-641f37cc4f6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.244146] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130477, 'name': ReconfigVM_Task, 'duration_secs': 0.292529} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.245960] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Reconfigured VM instance instance-00000072 to attach disk [datastore2] f2c347a3-47ef-48be-8aa6-1b45bc09e7c7/f2c347a3-47ef-48be-8aa6-1b45bc09e7c7.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.246604] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-088e0bef-c2ae-4984-8c92-92a57647bd33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.248733] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02051135-a102-433f-95c9-5de8ca653839 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.279015] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c321d989-b777-4097-b899-e43ca54227ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.282191] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1135.282191] env[61991]: value = "task-1130479" [ 1135.282191] env[61991]: _type = "Task" [ 1135.282191] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.289179] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef0e6a9-c548-49f8-ad55-5706f677fea5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.295757] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130479, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.306296] env[61991]: DEBUG nova.compute.provider_tree [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.514148] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130478, 'name': Rename_Task, 'duration_secs': 0.139269} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.514398] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.514638] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed7ced25-a535-44f6-a882-1e55322cfdf3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.520311] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1135.520311] env[61991]: value = "task-1130480" [ 1135.520311] env[61991]: _type = "Task" [ 1135.520311] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.527315] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130480, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.792831] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130479, 'name': Rename_Task, 'duration_secs': 0.147623} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.793198] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.793330] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b93048de-0d25-4108-bde5-13b64793fe73 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.799073] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1135.799073] env[61991]: value = "task-1130481" [ 1135.799073] env[61991]: _type = "Task" [ 1135.799073] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.806970] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130481, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.809103] env[61991]: DEBUG nova.scheduler.client.report [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.029904] env[61991]: DEBUG oslo_vmware.api [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130480, 'name': PowerOnVM_Task, 'duration_secs': 0.473876} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.030214] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.134830] env[61991]: DEBUG nova.compute.manager [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.135832] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1b089f-6ec7-4d7a-a0ea-a1b290c2fe1e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.311858] env[61991]: DEBUG oslo_vmware.api [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130481, 'name': PowerOnVM_Task, 'duration_secs': 0.50814} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.312187] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.312467] env[61991]: INFO nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Took 8.24 seconds to spawn the instance on the hypervisor. [ 1136.312725] env[61991]: DEBUG nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.313659] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.159s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.314303] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1136.318434] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaee635-83eb-4bf5-a46b-af2737a4a992 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.655027] env[61991]: DEBUG oslo_concurrency.lockutils [None req-ea222e6c-a008-4370-a420-0f6400bb5f9a tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.324s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.824130] env[61991]: DEBUG nova.compute.utils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1136.825635] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1136.825835] env[61991]: DEBUG nova.network.neutron [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1136.837864] env[61991]: INFO nova.compute.manager [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Took 16.65 seconds to build instance. [ 1136.864282] env[61991]: DEBUG nova.policy [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1137.178683] env[61991]: DEBUG nova.network.neutron [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Successfully created port: baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1137.239757] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.330557] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1137.342117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3435d00f-acac-4482-92ff-9fdfeb79b935 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.163s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.342117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.102s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.342512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.342762] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.342940] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.345124] env[61991]: INFO nova.compute.manager [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Terminating instance [ 1137.347025] env[61991]: DEBUG nova.compute.manager [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.347222] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.348082] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7436a18-9c50-417f-8f10-6320f86e47df {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.355885] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.356139] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97249c04-5d15-42fc-a583-42d67d0674ca {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.362090] env[61991]: DEBUG oslo_vmware.api [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1137.362090] env[61991]: value = "task-1130482" [ 1137.362090] env[61991]: _type = "Task" [ 1137.362090] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.370788] env[61991]: DEBUG oslo_vmware.api [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.799258] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.799360] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.799510] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "22f59842-073b-4bca-bf4f-a83552a90582-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.799714] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.799886] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.803054] env[61991]: INFO nova.compute.manager [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Terminating instance [ 1137.804638] env[61991]: DEBUG nova.compute.manager [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.804764] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.805625] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249a8594-5971-4cb2-a82d-00a554d865c8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.813087] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.813319] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-400038f8-e222-44c9-9c32-02687cf47722 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.819874] env[61991]: DEBUG oslo_vmware.api [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1137.819874] env[61991]: value = "task-1130483" [ 1137.819874] env[61991]: _type = "Task" [ 1137.819874] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.826748] env[61991]: DEBUG oslo_vmware.api [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.871733] env[61991]: DEBUG oslo_vmware.api [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130482, 'name': PowerOffVM_Task, 'duration_secs': 0.206712} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.872637] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1137.872897] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1137.873210] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e94dccb5-8f4c-4444-81ed-796dea528166 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.935921] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1137.936228] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1137.936414] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Deleting the datastore file [datastore2] f2c347a3-47ef-48be-8aa6-1b45bc09e7c7 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.936705] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8d57446-fcf8-4d93-a6bc-e0868f6f98bb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.943583] env[61991]: DEBUG oslo_vmware.api [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for the task: (returnval){ [ 1137.943583] env[61991]: value = "task-1130485" [ 1137.943583] env[61991]: _type = "Task" [ 1137.943583] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.951018] env[61991]: DEBUG oslo_vmware.api [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130485, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.330708] env[61991]: DEBUG oslo_vmware.api [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130483, 'name': PowerOffVM_Task, 'duration_secs': 0.193869} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.331016] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.331194] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1138.331447] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ed5b8bc-50ce-463a-a714-be4695d90d3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.340191] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1138.363821] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1138.364077] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1138.364243] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1138.364431] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1138.364583] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1138.364735] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1138.364943] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1138.365118] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1138.365301] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1138.365508] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1138.365682] env[61991]: DEBUG nova.virt.hardware [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1138.366776] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69825fd-8033-4353-8fce-279b8946e059 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.373968] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e479a0-b767-4139-9a38-a18deafc3119 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.399959] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1138.400237] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1138.400419] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleting the datastore file [datastore2] 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.400677] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-355b958d-216f-4552-ac96-54affba1ca58 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.408826] env[61991]: DEBUG oslo_vmware.api [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for the task: (returnval){ [ 1138.408826] env[61991]: value = "task-1130487" [ 1138.408826] env[61991]: _type = "Task" [ 1138.408826] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.419910] env[61991]: DEBUG oslo_vmware.api [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130487, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.454201] env[61991]: DEBUG oslo_vmware.api [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Task: {'id': task-1130485, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16005} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.454496] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.454726] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.454939] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.455175] env[61991]: INFO nova.compute.manager [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1138.455500] env[61991]: DEBUG oslo.service.loopingcall [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.455698] env[61991]: DEBUG nova.compute.manager [-] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.455815] env[61991]: DEBUG nova.network.neutron [-] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.876616] env[61991]: DEBUG nova.compute.manager [req-6092a78e-7792-4ef6-8fa9-5bde1d579617 req-aadbf880-01c3-4555-9fff-6f95787cfed2 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Received event network-vif-plugged-baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1138.876952] env[61991]: DEBUG oslo_concurrency.lockutils [req-6092a78e-7792-4ef6-8fa9-5bde1d579617 req-aadbf880-01c3-4555-9fff-6f95787cfed2 service nova] Acquiring lock "af7962c8-7700-4ff2-988e-65d2922fd514-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.877057] env[61991]: DEBUG oslo_concurrency.lockutils [req-6092a78e-7792-4ef6-8fa9-5bde1d579617 req-aadbf880-01c3-4555-9fff-6f95787cfed2 service nova] Lock "af7962c8-7700-4ff2-988e-65d2922fd514-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.877239] env[61991]: DEBUG oslo_concurrency.lockutils [req-6092a78e-7792-4ef6-8fa9-5bde1d579617 req-aadbf880-01c3-4555-9fff-6f95787cfed2 service nova] Lock "af7962c8-7700-4ff2-988e-65d2922fd514-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.877409] env[61991]: DEBUG nova.compute.manager [req-6092a78e-7792-4ef6-8fa9-5bde1d579617 req-aadbf880-01c3-4555-9fff-6f95787cfed2 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] No waiting events found dispatching network-vif-plugged-baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1138.877637] env[61991]: WARNING nova.compute.manager [req-6092a78e-7792-4ef6-8fa9-5bde1d579617 req-aadbf880-01c3-4555-9fff-6f95787cfed2 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Received unexpected event network-vif-plugged-baf8b828-c0f3-4cc5-8286-d92459511614 for instance with vm_state building and task_state spawning. [ 1138.919739] env[61991]: DEBUG oslo_vmware.api [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Task: {'id': task-1130487, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153032} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.919998] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.920880] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.921115] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.921331] env[61991]: INFO nova.compute.manager [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1138.928687] env[61991]: DEBUG oslo.service.loopingcall [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.928687] env[61991]: DEBUG nova.compute.manager [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.928687] env[61991]: DEBUG nova.network.neutron [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1139.088909] env[61991]: DEBUG nova.network.neutron [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Successfully updated port: baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1139.364148] env[61991]: DEBUG nova.compute.manager [req-d61f569a-9bac-423e-bc13-d708414a048b req-671cf37a-6a8a-4af7-8334-615437498595 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Received event network-vif-deleted-7d2cca63-6f27-4e38-a6b1-6e0eda1401e6 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.364578] env[61991]: INFO nova.compute.manager [req-d61f569a-9bac-423e-bc13-d708414a048b req-671cf37a-6a8a-4af7-8334-615437498595 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Neutron deleted interface 7d2cca63-6f27-4e38-a6b1-6e0eda1401e6; detaching it from the instance and deleting it from the info cache [ 1139.365489] env[61991]: DEBUG nova.network.neutron [req-d61f569a-9bac-423e-bc13-d708414a048b req-671cf37a-6a8a-4af7-8334-615437498595 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.555131] env[61991]: DEBUG nova.network.neutron [-] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.592385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-af7962c8-7700-4ff2-988e-65d2922fd514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.592385] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-af7962c8-7700-4ff2-988e-65d2922fd514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.592385] env[61991]: DEBUG nova.network.neutron [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1139.842275] env[61991]: DEBUG nova.network.neutron [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.867216] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acccb1f7-e6a1-471d-bc69-5ef66899aaaa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.877542] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e64397-e8fe-4a39-a2e4-b6728b22b526 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.904984] env[61991]: DEBUG nova.compute.manager [req-d61f569a-9bac-423e-bc13-d708414a048b req-671cf37a-6a8a-4af7-8334-615437498595 service nova] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Detach interface failed, port_id=7d2cca63-6f27-4e38-a6b1-6e0eda1401e6, reason: Instance 22f59842-073b-4bca-bf4f-a83552a90582 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1140.057721] env[61991]: INFO nova.compute.manager [-] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Took 1.60 seconds to deallocate network for instance. [ 1140.124535] env[61991]: DEBUG nova.network.neutron [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1140.261486] env[61991]: DEBUG nova.network.neutron [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Updating instance_info_cache with network_info: [{"id": "baf8b828-c0f3-4cc5-8286-d92459511614", "address": "fa:16:3e:5a:81:2a", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf8b828-c0", "ovs_interfaceid": "baf8b828-c0f3-4cc5-8286-d92459511614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.344642] env[61991]: INFO nova.compute.manager [-] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Took 1.42 seconds to deallocate network for instance. [ 1140.563947] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.564245] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.564473] env[61991]: DEBUG nova.objects.instance [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lazy-loading 'resources' on Instance uuid f2c347a3-47ef-48be-8aa6-1b45bc09e7c7 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.764815] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-af7962c8-7700-4ff2-988e-65d2922fd514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.765184] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Instance network_info: |[{"id": "baf8b828-c0f3-4cc5-8286-d92459511614", "address": "fa:16:3e:5a:81:2a", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf8b828-c0", "ovs_interfaceid": "baf8b828-c0f3-4cc5-8286-d92459511614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1140.765645] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:81:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'baf8b828-c0f3-4cc5-8286-d92459511614', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1140.773105] env[61991]: DEBUG oslo.service.loopingcall [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.773319] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1140.773547] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e629bb1d-ca29-4289-9c81-56d51e50d96f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.792988] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1140.792988] env[61991]: value = "task-1130488" [ 1140.792988] env[61991]: _type = "Task" [ 1140.792988] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.800372] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130488, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.851158] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.910505] env[61991]: DEBUG nova.compute.manager [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Received event network-vif-deleted-e0b63cd0-e600-4bd2-ad7d-e1a1c450acc5 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.910792] env[61991]: DEBUG nova.compute.manager [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Received event network-changed-baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1140.911017] env[61991]: DEBUG nova.compute.manager [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Refreshing instance network info cache due to event network-changed-baf8b828-c0f3-4cc5-8286-d92459511614. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1140.911455] env[61991]: DEBUG oslo_concurrency.lockutils [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] Acquiring lock "refresh_cache-af7962c8-7700-4ff2-988e-65d2922fd514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.911698] env[61991]: DEBUG oslo_concurrency.lockutils [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] Acquired lock "refresh_cache-af7962c8-7700-4ff2-988e-65d2922fd514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.911899] env[61991]: DEBUG nova.network.neutron [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Refreshing network info cache for port baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.141033] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25adcec5-67f5-4cc6-932d-0d251ee969f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.150014] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb5deb0-de92-42aa-8c76-abbd69c3cf1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.177813] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b7a98f-12ed-4dcf-969a-75531d65a8c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.184358] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86a477e-605d-44b5-b529-7b43ff04656e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.197985] env[61991]: DEBUG nova.compute.provider_tree [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.302628] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130488, 'name': CreateVM_Task, 'duration_secs': 0.297715} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.302806] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1141.303559] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.303736] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.304069] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1141.304318] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8da6e4d4-0e0a-48c1-b3b3-cd5e3790ad48 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.308412] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1141.308412] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52484228-7151-3751-b561-3e156e354f5b" [ 1141.308412] env[61991]: _type = "Task" [ 1141.308412] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.315506] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52484228-7151-3751-b561-3e156e354f5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.585844] env[61991]: DEBUG nova.network.neutron [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Updated VIF entry in instance network info cache for port baf8b828-c0f3-4cc5-8286-d92459511614. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.586229] env[61991]: DEBUG nova.network.neutron [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Updating instance_info_cache with network_info: [{"id": "baf8b828-c0f3-4cc5-8286-d92459511614", "address": "fa:16:3e:5a:81:2a", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbaf8b828-c0", "ovs_interfaceid": "baf8b828-c0f3-4cc5-8286-d92459511614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.701406] env[61991]: DEBUG nova.scheduler.client.report [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.818764] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52484228-7151-3751-b561-3e156e354f5b, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.819117] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.819357] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.819603] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.819766] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.819949] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.820234] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b8295b5-bd83-4fde-a6bb-b71f933bfbd4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.828309] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.828488] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1141.829209] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85b09b3e-9668-493b-af45-b8c09ba5cfcf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.834175] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1141.834175] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c73f4-8d07-d13e-5a09-d36e73a248cd" [ 1141.834175] env[61991]: _type = "Task" [ 1141.834175] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.841473] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c73f4-8d07-d13e-5a09-d36e73a248cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.088934] env[61991]: DEBUG oslo_concurrency.lockutils [req-75d84d2a-6dbb-46e5-8d99-5873a70a0b9f req-9c0fc1d0-c875-4749-bfe3-a180a2739e63 service nova] Releasing lock "refresh_cache-af7962c8-7700-4ff2-988e-65d2922fd514" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.206380] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.642s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.208861] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.358s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.209121] env[61991]: DEBUG nova.objects.instance [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lazy-loading 'resources' on Instance uuid 22f59842-073b-4bca-bf4f-a83552a90582 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.227731] env[61991]: INFO nova.scheduler.client.report [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Deleted allocations for instance f2c347a3-47ef-48be-8aa6-1b45bc09e7c7 [ 1142.344631] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]526c73f4-8d07-d13e-5a09-d36e73a248cd, 'name': SearchDatastore_Task, 'duration_secs': 0.008624} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.345442] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f22a38e-9ab8-4da4-a3b6-f8ffc08cf691 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.350277] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1142.350277] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5213eb11-3e2b-a33f-f0ab-4334b5a86c8a" [ 1142.350277] env[61991]: _type = "Task" [ 1142.350277] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.357740] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5213eb11-3e2b-a33f-f0ab-4334b5a86c8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.734846] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b2481878-9163-4e33-87df-554efe453c61 tempest-ServerAddressesTestJSON-1046558726 tempest-ServerAddressesTestJSON-1046558726-project-member] Lock "f2c347a3-47ef-48be-8aa6-1b45bc09e7c7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.393s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.765385] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f76e3b7-4d17-4d87-9b8f-a2e3d9e15280 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.773227] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6881d368-9aef-41a3-8039-90b38ade8a56 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.804862] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c276d711-26b2-498c-835d-0cf81cc9ca55 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.811096] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e97197-d7bd-4c38-9404-821af154cbdc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.823537] env[61991]: DEBUG nova.compute.provider_tree [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.859299] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5213eb11-3e2b-a33f-f0ab-4334b5a86c8a, 'name': SearchDatastore_Task, 'duration_secs': 0.010068} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.859556] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.859801] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] af7962c8-7700-4ff2-988e-65d2922fd514/af7962c8-7700-4ff2-988e-65d2922fd514.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1142.860074] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1860d0ad-f3fa-4ce5-9e3a-b0faaa76dad2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.866354] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1142.866354] env[61991]: value = "task-1130489" [ 1142.866354] env[61991]: _type = "Task" [ 1142.866354] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.873425] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130489, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.326706] env[61991]: DEBUG nova.scheduler.client.report [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.376282] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130489, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478158} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.376540] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] af7962c8-7700-4ff2-988e-65d2922fd514/af7962c8-7700-4ff2-988e-65d2922fd514.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1143.376754] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1143.376996] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3777c580-7f62-456c-8687-edb0a465063f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.383303] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1143.383303] env[61991]: value = "task-1130490" [ 1143.383303] env[61991]: _type = "Task" [ 1143.383303] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.390504] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.832234] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.855165] env[61991]: INFO nova.scheduler.client.report [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Deleted allocations for instance 22f59842-073b-4bca-bf4f-a83552a90582 [ 1143.898101] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091363} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.898348] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1143.899308] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7085dd11-9cec-4974-8dab-d8ada8ab941f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.922322] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] af7962c8-7700-4ff2-988e-65d2922fd514/af7962c8-7700-4ff2-988e-65d2922fd514.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1143.922322] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ebc5c1f-1a8b-400a-b449-9618c2d5b870 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.942101] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1143.942101] env[61991]: value = "task-1130491" [ 1143.942101] env[61991]: _type = "Task" [ 1143.942101] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.952717] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130491, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.363499] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e3c8b8eb-9ef3-4b82-930e-6b4acf3920e1 tempest-ServerActionsTestOtherB-40966218 tempest-ServerActionsTestOtherB-40966218-project-member] Lock "22f59842-073b-4bca-bf4f-a83552a90582" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.564s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.455047] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130491, 'name': ReconfigVM_Task, 'duration_secs': 0.273116} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.456462] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Reconfigured VM instance instance-00000073 to attach disk [datastore1] af7962c8-7700-4ff2-988e-65d2922fd514/af7962c8-7700-4ff2-988e-65d2922fd514.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.457351] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9433bf7f-01ec-417c-bf5a-2a363cbc5965 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.463772] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1144.463772] env[61991]: value = "task-1130493" [ 1144.463772] env[61991]: _type = "Task" [ 1144.463772] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.472405] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130493, 'name': Rename_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.973714] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130493, 'name': Rename_Task, 'duration_secs': 0.13738} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.973888] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1144.974955] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29cabcc5-8ce4-4c2c-a198-2beaecc9ed40 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.981294] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1144.981294] env[61991]: value = "task-1130494" [ 1144.981294] env[61991]: _type = "Task" [ 1144.981294] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.988589] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130494, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.491509] env[61991]: DEBUG oslo_vmware.api [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130494, 'name': PowerOnVM_Task, 'duration_secs': 0.428476} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.491874] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1145.491974] env[61991]: INFO nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1145.492182] env[61991]: DEBUG nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1145.492911] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ebe9d5-8974-4dae-b8de-5ea5efd395e1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.009807] env[61991]: INFO nova.compute.manager [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Took 11.87 seconds to build instance. [ 1146.512336] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e5d1a8b7-09e7-49f0-a0aa-fa8520d1648e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "af7962c8-7700-4ff2-988e-65d2922fd514" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.384s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.649551] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "af7962c8-7700-4ff2-988e-65d2922fd514" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.649834] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "af7962c8-7700-4ff2-988e-65d2922fd514" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.650068] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "af7962c8-7700-4ff2-988e-65d2922fd514-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.650279] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "af7962c8-7700-4ff2-988e-65d2922fd514-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.650454] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "af7962c8-7700-4ff2-988e-65d2922fd514-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.652682] env[61991]: INFO nova.compute.manager [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Terminating instance [ 1146.654439] env[61991]: DEBUG nova.compute.manager [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1146.654634] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1146.655459] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7edd96c-986a-4de9-8333-b27c74f9c1ad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.663054] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.663282] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2546b3b-5e9b-4631-8827-8aee30a67be1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.669585] env[61991]: DEBUG oslo_vmware.api [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1146.669585] env[61991]: value = "task-1130495" [ 1146.669585] env[61991]: _type = "Task" [ 1146.669585] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.676797] env[61991]: DEBUG oslo_vmware.api [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.180128] env[61991]: DEBUG oslo_vmware.api [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130495, 'name': PowerOffVM_Task, 'duration_secs': 0.182989} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.180405] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1147.180574] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1147.180815] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4302120e-8863-40a5-b1a8-d8b42a7cb8fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.240817] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1147.241090] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1147.241283] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore1] af7962c8-7700-4ff2-988e-65d2922fd514 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1147.241543] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5d8c0c1-a817-477a-a5bd-35c3fcda5f44 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.247912] env[61991]: DEBUG oslo_vmware.api [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1147.247912] env[61991]: value = "task-1130497" [ 1147.247912] env[61991]: _type = "Task" [ 1147.247912] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.255492] env[61991]: DEBUG oslo_vmware.api [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.760440] env[61991]: DEBUG oslo_vmware.api [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230702} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.760440] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1147.760440] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1147.760440] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1147.760440] env[61991]: INFO nova.compute.manager [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1147.760440] env[61991]: DEBUG oslo.service.loopingcall [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.760440] env[61991]: DEBUG nova.compute.manager [-] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1147.760440] env[61991]: DEBUG nova.network.neutron [-] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1147.999231] env[61991]: DEBUG nova.compute.manager [req-d6d92740-3e07-478d-8865-da1d1035b86a req-3e867369-4aa3-4942-9b89-377ef608227e service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Received event network-vif-deleted-baf8b828-c0f3-4cc5-8286-d92459511614 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.999429] env[61991]: INFO nova.compute.manager [req-d6d92740-3e07-478d-8865-da1d1035b86a req-3e867369-4aa3-4942-9b89-377ef608227e service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Neutron deleted interface baf8b828-c0f3-4cc5-8286-d92459511614; detaching it from the instance and deleting it from the info cache [ 1147.999611] env[61991]: DEBUG nova.network.neutron [req-d6d92740-3e07-478d-8865-da1d1035b86a req-3e867369-4aa3-4942-9b89-377ef608227e service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.464504] env[61991]: DEBUG nova.network.neutron [-] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.502609] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f57850f3-fdad-4944-894e-368ffc83f532 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.513223] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f399afa-062c-4b04-8d05-7e3d23e74b52 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.538332] env[61991]: DEBUG nova.compute.manager [req-d6d92740-3e07-478d-8865-da1d1035b86a req-3e867369-4aa3-4942-9b89-377ef608227e service nova] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Detach interface failed, port_id=baf8b828-c0f3-4cc5-8286-d92459511614, reason: Instance af7962c8-7700-4ff2-988e-65d2922fd514 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1148.967162] env[61991]: INFO nova.compute.manager [-] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Took 1.21 seconds to deallocate network for instance. [ 1149.474296] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.474598] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.474825] env[61991]: DEBUG nova.objects.instance [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'resources' on Instance uuid af7962c8-7700-4ff2-988e-65d2922fd514 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1150.035731] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58301c3-03a3-4c8d-aeff-0dd49989c7f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.043664] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38acb6f0-4b78-42ce-be1b-44c5b189f55c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.076803] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3be82b9-7657-42b5-a6c6-8adb83b6c75f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.084761] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eab8265-d8f9-4bbf-a76e-83be4eca9f80 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.099290] env[61991]: DEBUG nova.compute.provider_tree [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.600350] env[61991]: DEBUG nova.scheduler.client.report [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1151.107891] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.135842] env[61991]: INFO nova.scheduler.client.report [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance af7962c8-7700-4ff2-988e-65d2922fd514 [ 1151.647318] env[61991]: DEBUG oslo_concurrency.lockutils [None req-432528eb-5a32-4eef-9639-4987dcd6544c tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "af7962c8-7700-4ff2-988e-65d2922fd514" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.997s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.599708] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "6a90ba3f-b947-4f30-8699-29537b97a214" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.600071] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.102173] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1153.627267] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.627579] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.629198] env[61991]: INFO nova.compute.claims [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.688983] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf12f8f-9f96-473f-b722-17099a63f31b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.698157] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ced1d4-4a12-4096-9944-bde2beaf7e46 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.727460] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dcd92d-851f-4cd8-8bde-6b433c402c89 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.734300] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f428db-3a65-4699-8f64-18ff84e8cb4c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.747028] env[61991]: DEBUG nova.compute.provider_tree [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.250113] env[61991]: DEBUG nova.scheduler.client.report [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.754903] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.127s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.755539] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1156.259978] env[61991]: DEBUG nova.compute.utils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.261461] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1156.261660] env[61991]: DEBUG nova.network.neutron [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1156.307925] env[61991]: DEBUG nova.policy [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d9bba4d7da4cd69792d9e49a55db71', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07a063bdfecb4cf7956b078c86f239f2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1156.708891] env[61991]: DEBUG nova.network.neutron [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Successfully created port: 9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1156.764726] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1157.776877] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1157.801926] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1157.802209] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1157.802373] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.802560] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1157.802710] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.802857] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1157.803079] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1157.803252] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1157.803424] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1157.803586] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1157.803758] env[61991]: DEBUG nova.virt.hardware [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1157.804679] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e146e4e-be32-4ea8-9505-beb043ef2c62 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.812696] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506756c8-7976-4f60-a14d-5acc3b9a4281 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.066433] env[61991]: DEBUG nova.compute.manager [req-2fcf32f3-585b-4814-83ae-57350c89479d req-b6bb7dab-a7a6-4b31-b32c-18669b8ebb92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Received event network-vif-plugged-9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.066639] env[61991]: DEBUG oslo_concurrency.lockutils [req-2fcf32f3-585b-4814-83ae-57350c89479d req-b6bb7dab-a7a6-4b31-b32c-18669b8ebb92 service nova] Acquiring lock "6a90ba3f-b947-4f30-8699-29537b97a214-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.066850] env[61991]: DEBUG oslo_concurrency.lockutils [req-2fcf32f3-585b-4814-83ae-57350c89479d req-b6bb7dab-a7a6-4b31-b32c-18669b8ebb92 service nova] Lock "6a90ba3f-b947-4f30-8699-29537b97a214-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.067024] env[61991]: DEBUG oslo_concurrency.lockutils [req-2fcf32f3-585b-4814-83ae-57350c89479d req-b6bb7dab-a7a6-4b31-b32c-18669b8ebb92 service nova] Lock "6a90ba3f-b947-4f30-8699-29537b97a214-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.067193] env[61991]: DEBUG nova.compute.manager [req-2fcf32f3-585b-4814-83ae-57350c89479d req-b6bb7dab-a7a6-4b31-b32c-18669b8ebb92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] No waiting events found dispatching network-vif-plugged-9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1158.067354] env[61991]: WARNING nova.compute.manager [req-2fcf32f3-585b-4814-83ae-57350c89479d req-b6bb7dab-a7a6-4b31-b32c-18669b8ebb92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Received unexpected event network-vif-plugged-9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 for instance with vm_state building and task_state spawning. [ 1158.148308] env[61991]: DEBUG nova.network.neutron [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Successfully updated port: 9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1158.650939] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "refresh_cache-6a90ba3f-b947-4f30-8699-29537b97a214" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.651134] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "refresh_cache-6a90ba3f-b947-4f30-8699-29537b97a214" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.651358] env[61991]: DEBUG nova.network.neutron [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1159.181419] env[61991]: DEBUG nova.network.neutron [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1159.303971] env[61991]: DEBUG nova.network.neutron [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Updating instance_info_cache with network_info: [{"id": "9b0e1e60-dc43-40ad-8b7b-3598b30e07e7", "address": "fa:16:3e:29:d0:a2", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0e1e60-dc", "ovs_interfaceid": "9b0e1e60-dc43-40ad-8b7b-3598b30e07e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.806932] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "refresh_cache-6a90ba3f-b947-4f30-8699-29537b97a214" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.807257] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Instance network_info: |[{"id": "9b0e1e60-dc43-40ad-8b7b-3598b30e07e7", "address": "fa:16:3e:29:d0:a2", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0e1e60-dc", "ovs_interfaceid": "9b0e1e60-dc43-40ad-8b7b-3598b30e07e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1159.807736] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:d0:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b0e1e60-dc43-40ad-8b7b-3598b30e07e7', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1159.815255] env[61991]: DEBUG oslo.service.loopingcall [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.815475] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1159.815703] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5de6dc2-bdc2-4bb0-9425-8a69d5188b6b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.835395] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1159.835395] env[61991]: value = "task-1130504" [ 1159.835395] env[61991]: _type = "Task" [ 1159.835395] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.842413] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130504, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.098834] env[61991]: DEBUG nova.compute.manager [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Received event network-changed-9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1160.099093] env[61991]: DEBUG nova.compute.manager [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Refreshing instance network info cache due to event network-changed-9b0e1e60-dc43-40ad-8b7b-3598b30e07e7. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1160.099325] env[61991]: DEBUG oslo_concurrency.lockutils [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] Acquiring lock "refresh_cache-6a90ba3f-b947-4f30-8699-29537b97a214" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.099474] env[61991]: DEBUG oslo_concurrency.lockutils [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] Acquired lock "refresh_cache-6a90ba3f-b947-4f30-8699-29537b97a214" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.099643] env[61991]: DEBUG nova.network.neutron [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Refreshing network info cache for port 9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1160.345856] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130504, 'name': CreateVM_Task, 'duration_secs': 0.453572} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.346143] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1160.346756] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.346944] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.347286] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1160.347538] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da81194e-deeb-42b8-9b5b-798e1c9a626f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.352975] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1160.352975] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b4db47-1c43-6fe2-092b-d342635304be" [ 1160.352975] env[61991]: _type = "Task" [ 1160.352975] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.361192] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b4db47-1c43-6fe2-092b-d342635304be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.783498] env[61991]: DEBUG nova.network.neutron [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Updated VIF entry in instance network info cache for port 9b0e1e60-dc43-40ad-8b7b-3598b30e07e7. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1160.783861] env[61991]: DEBUG nova.network.neutron [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Updating instance_info_cache with network_info: [{"id": "9b0e1e60-dc43-40ad-8b7b-3598b30e07e7", "address": "fa:16:3e:29:d0:a2", "network": {"id": "0df7776c-45b2-4a3f-a180-4acd9b0cc182", "bridge": "br-int", "label": "tempest-ServersTestJSON-617087169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07a063bdfecb4cf7956b078c86f239f2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b0e1e60-dc", "ovs_interfaceid": "9b0e1e60-dc43-40ad-8b7b-3598b30e07e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.862550] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52b4db47-1c43-6fe2-092b-d342635304be, 'name': SearchDatastore_Task, 'duration_secs': 0.040265} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.862859] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.863110] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.863361] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.863512] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.863750] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.864052] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e54c33b-69ef-4029-92ba-1e0df08f67d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.872818] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.873019] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1160.873737] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-624465cb-c532-4e45-b40b-a2402f44e96a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.878597] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1160.878597] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520283fa-5071-fc4b-1c9f-290d266b3c3b" [ 1160.878597] env[61991]: _type = "Task" [ 1160.878597] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.885854] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520283fa-5071-fc4b-1c9f-290d266b3c3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.287827] env[61991]: DEBUG oslo_concurrency.lockutils [req-fec3811d-ad7c-47d5-9324-d467dd69b2d2 req-3edf5e68-cd87-4419-a9e5-994ecb48d883 service nova] Releasing lock "refresh_cache-6a90ba3f-b947-4f30-8699-29537b97a214" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.388328] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]520283fa-5071-fc4b-1c9f-290d266b3c3b, 'name': SearchDatastore_Task, 'duration_secs': 0.008245} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.389183] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16738c50-66cc-4772-91c4-3f94e1123ce4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.394551] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1161.394551] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bf23cc-4d31-ac8e-07ec-9b2f01b9197b" [ 1161.394551] env[61991]: _type = "Task" [ 1161.394551] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.402914] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bf23cc-4d31-ac8e-07ec-9b2f01b9197b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.904748] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52bf23cc-4d31-ac8e-07ec-9b2f01b9197b, 'name': SearchDatastore_Task, 'duration_secs': 0.009177} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.905165] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.905455] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 6a90ba3f-b947-4f30-8699-29537b97a214/6a90ba3f-b947-4f30-8699-29537b97a214.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1161.905718] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad2b8aa6-a3c0-41d6-b759-9258e12c2369 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.911797] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1161.911797] env[61991]: value = "task-1130506" [ 1161.911797] env[61991]: _type = "Task" [ 1161.911797] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.919357] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.421411] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447127} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.421677] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 6a90ba3f-b947-4f30-8699-29537b97a214/6a90ba3f-b947-4f30-8699-29537b97a214.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1162.421906] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.422198] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71b06dd1-b4c2-4400-9f5f-b0d11cb9a153 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.427548] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1162.427548] env[61991]: value = "task-1130507" [ 1162.427548] env[61991]: _type = "Task" [ 1162.427548] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.434053] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.938053] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131813} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.938053] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1162.938681] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26781b3c-bd74-4dc5-b0e3-98ae4df7f447 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.959849] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 6a90ba3f-b947-4f30-8699-29537b97a214/6a90ba3f-b947-4f30-8699-29537b97a214.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.960116] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82874665-ef37-47ac-8e0f-cd3931e7bf6f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.979238] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1162.979238] env[61991]: value = "task-1130508" [ 1162.979238] env[61991]: _type = "Task" [ 1162.979238] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.987868] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.489267] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130508, 'name': ReconfigVM_Task, 'duration_secs': 0.280655} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.489595] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 6a90ba3f-b947-4f30-8699-29537b97a214/6a90ba3f-b947-4f30-8699-29537b97a214.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.490153] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf3873e8-0829-4a71-9600-8e85e1da74cb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.495652] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1163.495652] env[61991]: value = "task-1130510" [ 1163.495652] env[61991]: _type = "Task" [ 1163.495652] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.503729] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130510, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.005593] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130510, 'name': Rename_Task, 'duration_secs': 0.154115} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.005904] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1164.006192] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58dc7055-414e-4bbf-abf9-a750dcc7c014 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.011910] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1164.011910] env[61991]: value = "task-1130511" [ 1164.011910] env[61991]: _type = "Task" [ 1164.011910] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.018885] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130511, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.522164] env[61991]: DEBUG oslo_vmware.api [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130511, 'name': PowerOnVM_Task, 'duration_secs': 0.402931} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.522527] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1164.522626] env[61991]: INFO nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Took 6.75 seconds to spawn the instance on the hypervisor. [ 1164.522803] env[61991]: DEBUG nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1164.523591] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa352517-4133-4553-ad6b-a3c57b837048 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.041132] env[61991]: INFO nova.compute.manager [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Took 11.43 seconds to build instance. [ 1165.543699] env[61991]: DEBUG oslo_concurrency.lockutils [None req-f0e8756a-654f-4e16-87bb-3853a8ac61aa tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.943s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.596860] env[61991]: DEBUG oslo_concurrency.lockutils [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "f42c1c55-3635-460e-a106-e63e57426fd2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.597107] env[61991]: DEBUG oslo_concurrency.lockutils [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.597348] env[61991]: DEBUG nova.compute.manager [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1165.598250] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6156d9c5-6e67-45fe-a8b3-973ea9ba1b2b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.605300] env[61991]: DEBUG nova.compute.manager [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1165.605897] env[61991]: DEBUG nova.objects.instance [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'flavor' on Instance uuid f42c1c55-3635-460e-a106-e63e57426fd2 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.012363] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "6a90ba3f-b947-4f30-8699-29537b97a214" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.012781] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.013072] env[61991]: DEBUG nova.compute.manager [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.014351] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695c7f8a-d345-4105-ad24-4cd56eec931c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.022689] env[61991]: DEBUG nova.compute.manager [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1166.023464] env[61991]: DEBUG nova.objects.instance [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'flavor' on Instance uuid 6a90ba3f-b947-4f30-8699-29537b97a214 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.110810] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1166.111103] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5bc6aec2-62c4-470e-b81b-03aa5cbc7b5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.119147] env[61991]: DEBUG oslo_vmware.api [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1166.119147] env[61991]: value = "task-1130513" [ 1166.119147] env[61991]: _type = "Task" [ 1166.119147] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.127222] env[61991]: DEBUG oslo_vmware.api [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130513, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.528918] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1166.529246] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bee2070-bc01-46a6-acc5-aba6a66b5e76 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.535811] env[61991]: DEBUG oslo_vmware.api [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1166.535811] env[61991]: value = "task-1130514" [ 1166.535811] env[61991]: _type = "Task" [ 1166.535811] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.543616] env[61991]: DEBUG oslo_vmware.api [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.630621] env[61991]: DEBUG oslo_vmware.api [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130513, 'name': PowerOffVM_Task, 'duration_secs': 0.179988} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.630997] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1166.631280] env[61991]: DEBUG nova.compute.manager [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.632473] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eba23ad-c6f8-45f2-99a1-82b405ba7443 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.046600] env[61991]: DEBUG oslo_vmware.api [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130514, 'name': PowerOffVM_Task, 'duration_secs': 0.18273} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.046867] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1167.047057] env[61991]: DEBUG nova.compute.manager [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1167.047818] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fac69e3-d4e8-4d99-84a1-d11f8628b462 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.146493] env[61991]: DEBUG oslo_concurrency.lockutils [None req-059e0f68-87b9-45dc-a179-fe48f6e59552 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.559450] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6240b377-791c-4391-a758-2e4946c33397 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.013075] env[61991]: DEBUG nova.objects.instance [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'flavor' on Instance uuid f42c1c55-3635-460e-a106-e63e57426fd2 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.518339] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.518527] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.518682] env[61991]: DEBUG nova.network.neutron [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1168.518898] env[61991]: DEBUG nova.objects.instance [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'info_cache' on Instance uuid f42c1c55-3635-460e-a106-e63e57426fd2 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.634673] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "6a90ba3f-b947-4f30-8699-29537b97a214" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.634949] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.635188] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "6a90ba3f-b947-4f30-8699-29537b97a214-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.635379] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.635550] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.637811] env[61991]: INFO nova.compute.manager [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Terminating instance [ 1168.639684] env[61991]: DEBUG nova.compute.manager [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1168.639881] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1168.640811] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0eb2275-bd02-443b-89f7-a06820eb1d33 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.649491] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1168.649735] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f023206b-302c-4f23-8f77-1d7f962da0f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.722969] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1168.723223] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1168.723413] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore1] 6a90ba3f-b947-4f30-8699-29537b97a214 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1168.723675] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c6a3f48-a43f-4bb9-b701-27ef4c0f1c03 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.730540] env[61991]: DEBUG oslo_vmware.api [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1168.730540] env[61991]: value = "task-1130517" [ 1168.730540] env[61991]: _type = "Task" [ 1168.730540] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.738101] env[61991]: DEBUG oslo_vmware.api [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130517, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.022873] env[61991]: DEBUG nova.objects.base [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1169.239817] env[61991]: DEBUG oslo_vmware.api [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130517, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.117786} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.240156] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1169.240361] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1169.240601] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1169.240817] env[61991]: INFO nova.compute.manager [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1169.241073] env[61991]: DEBUG oslo.service.loopingcall [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1169.241346] env[61991]: DEBUG nova.compute.manager [-] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1169.241469] env[61991]: DEBUG nova.network.neutron [-] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1169.496268] env[61991]: DEBUG nova.compute.manager [req-eab8f495-75c8-4e6e-9f01-3cd62a7e4fa2 req-910d61a6-402b-413e-9e46-2d2f9c000f92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Received event network-vif-deleted-9b0e1e60-dc43-40ad-8b7b-3598b30e07e7 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1169.496511] env[61991]: INFO nova.compute.manager [req-eab8f495-75c8-4e6e-9f01-3cd62a7e4fa2 req-910d61a6-402b-413e-9e46-2d2f9c000f92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Neutron deleted interface 9b0e1e60-dc43-40ad-8b7b-3598b30e07e7; detaching it from the instance and deleting it from the info cache [ 1169.496665] env[61991]: DEBUG nova.network.neutron [req-eab8f495-75c8-4e6e-9f01-3cd62a7e4fa2 req-910d61a6-402b-413e-9e46-2d2f9c000f92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.721255] env[61991]: DEBUG nova.network.neutron [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [{"id": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "address": "fa:16:3e:c0:1c:8c", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5f87d87-c6", "ovs_interfaceid": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.972810] env[61991]: DEBUG nova.network.neutron [-] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.999771] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3a5f58c-3e47-4234-83f5-1e7fd8397210 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.010305] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd639fd-5783-4408-99cd-21bdacd9ec0c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.034829] env[61991]: DEBUG nova.compute.manager [req-eab8f495-75c8-4e6e-9f01-3cd62a7e4fa2 req-910d61a6-402b-413e-9e46-2d2f9c000f92 service nova] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Detach interface failed, port_id=9b0e1e60-dc43-40ad-8b7b-3598b30e07e7, reason: Instance 6a90ba3f-b947-4f30-8699-29537b97a214 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1170.224153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.476137] env[61991]: INFO nova.compute.manager [-] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Took 1.23 seconds to deallocate network for instance. [ 1170.631864] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.632123] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.727383] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1170.727667] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b6f0603-d38d-44c9-839d-04d44df6d40c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.735231] env[61991]: DEBUG oslo_vmware.api [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1170.735231] env[61991]: value = "task-1130518" [ 1170.735231] env[61991]: _type = "Task" [ 1170.735231] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.742744] env[61991]: DEBUG oslo_vmware.api [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130518, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.983016] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.983430] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.983670] env[61991]: DEBUG nova.objects.instance [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'resources' on Instance uuid 6a90ba3f-b947-4f30-8699-29537b97a214 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.134452] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1171.244973] env[61991]: DEBUG oslo_vmware.api [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130518, 'name': PowerOnVM_Task, 'duration_secs': 0.354495} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.245282] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1171.245453] env[61991]: DEBUG nova.compute.manager [None req-6ba0b4b6-b030-4317-93c6-6ece791dc022 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1171.246239] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15417c94-62b2-4659-9d7e-16d3ccb87543 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.544584] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa67ca8-9251-4b5b-abb2-2b5dc6cf97c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.552136] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954a2218-af93-49d4-8721-0334cecabb88 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.581838] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917f7d32-d728-41bf-986d-54d5bef1d5b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.588302] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af058a8-53df-456a-a493-20b512227161 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.601022] env[61991]: DEBUG nova.compute.provider_tree [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.652670] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.969891] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.103814] env[61991]: DEBUG nova.scheduler.client.report [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1172.609279] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.612019] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.959s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.613965] env[61991]: INFO nova.compute.claims [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1172.627243] env[61991]: INFO nova.scheduler.client.report [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance 6a90ba3f-b947-4f30-8699-29537b97a214 [ 1172.969903] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.970137] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.065896] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6501e9-1029-4dbc-b0dc-35ff75011e2a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.072942] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Suspending the VM {{(pid=61991) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1173.073203] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-415eaf74-7968-448f-85be-c75b4a99a650 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.079516] env[61991]: DEBUG oslo_vmware.api [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1173.079516] env[61991]: value = "task-1130519" [ 1173.079516] env[61991]: _type = "Task" [ 1173.079516] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.086784] env[61991]: DEBUG oslo_vmware.api [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130519, 'name': SuspendVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.134681] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b7a06bd3-8578-4852-8b28-f30585cc774e tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "6a90ba3f-b947-4f30-8699-29537b97a214" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.500s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.589532] env[61991]: DEBUG oslo_vmware.api [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130519, 'name': SuspendVM_Task} progress is 66%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.683803] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9599bf-bb68-4b45-bffe-a3f864fa1324 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.691070] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6ec07f-3819-4c39-a0df-b6dfbad7cb37 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.721810] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d57820-e67d-4763-8c6e-151a42589b67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.729849] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3e128e-632a-41e3-8417-dc1f3101a940 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.743189] env[61991]: DEBUG nova.compute.provider_tree [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.969422] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.969605] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1174.088977] env[61991]: DEBUG oslo_vmware.api [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130519, 'name': SuspendVM_Task, 'duration_secs': 0.599567} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.089231] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Suspended the VM {{(pid=61991) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1174.089413] env[61991]: DEBUG nova.compute.manager [None req-f0118fe7-4bdf-47ac-84ef-2fe417492e14 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1174.090174] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bb99b8-2f9c-4b17-b9b9-b6ac47c040ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.247118] env[61991]: DEBUG nova.scheduler.client.report [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1174.387294] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.387560] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.387773] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.388030] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.388217] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.390619] env[61991]: INFO nova.compute.manager [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Terminating instance [ 1174.392257] env[61991]: DEBUG nova.compute.manager [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1174.392451] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.393500] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923fc7ed-c0d9-4964-9a3f-a631cc1dc32a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.400787] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1174.400984] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7288441-3e54-4bb9-ac8f-ae6d633b8aa9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.406742] env[61991]: DEBUG oslo_vmware.api [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1174.406742] env[61991]: value = "task-1130520" [ 1174.406742] env[61991]: _type = "Task" [ 1174.406742] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.415654] env[61991]: DEBUG oslo_vmware.api [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.752025] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.140s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.752434] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1174.917073] env[61991]: DEBUG oslo_vmware.api [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130520, 'name': PowerOffVM_Task, 'duration_secs': 0.163213} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.917366] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.917538] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.917799] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6abaf795-4665-4e23-9851-e5a891164347 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.969846] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1174.974427] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.974643] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.974821] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleting the datastore file [datastore1] 84a0dee7-c33b-494a-ad38-83da6ab44ce2 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.975078] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14856373-7df1-49d0-b909-515a78e0a852 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.981912] env[61991]: DEBUG oslo_vmware.api [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for the task: (returnval){ [ 1174.981912] env[61991]: value = "task-1130522" [ 1174.981912] env[61991]: _type = "Task" [ 1174.981912] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.990751] env[61991]: DEBUG oslo_vmware.api [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.257370] env[61991]: DEBUG nova.compute.utils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1175.259273] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1175.259444] env[61991]: DEBUG nova.network.neutron [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1175.296736] env[61991]: DEBUG nova.policy [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '049ce8091c454bbdac1e4e0f7be19c5d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97761803e1374ac2a1d9028c79cbccde', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1175.440640] env[61991]: INFO nova.compute.manager [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Resuming [ 1175.441308] env[61991]: DEBUG nova.objects.instance [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'flavor' on Instance uuid f42c1c55-3635-460e-a106-e63e57426fd2 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.472818] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.473057] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.473228] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.473379] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1175.474488] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b98a88-f1f3-4790-9e24-fc4f0aa74ed3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.482656] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668365d8-1043-4fc2-9925-c76fb8b4f1e5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.494342] env[61991]: DEBUG oslo_vmware.api [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Task: {'id': task-1130522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131206} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.501990] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.501990] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.501990] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.502124] env[61991]: INFO nova.compute.manager [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1175.502404] env[61991]: DEBUG oslo.service.loopingcall [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1175.502630] env[61991]: DEBUG nova.compute.manager [-] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1175.502728] env[61991]: DEBUG nova.network.neutron [-] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1175.504703] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9894f673-ef83-4fdd-9c0a-93b30997df5e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.511155] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678ca97e-8822-4873-856e-b84ba196edb8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.541283] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181187MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1175.541443] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.541645] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.586271] env[61991]: DEBUG nova.network.neutron [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Successfully created port: e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1175.738284] env[61991]: DEBUG nova.compute.manager [req-a202ab54-b25b-4ed6-bc9f-eee14029d954 req-4a9384f9-a680-440d-8e4a-69609bf28a9c service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Received event network-vif-deleted-7f4f6f32-7ee7-4884-86c4-27ecf7daca85 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1175.738525] env[61991]: INFO nova.compute.manager [req-a202ab54-b25b-4ed6-bc9f-eee14029d954 req-4a9384f9-a680-440d-8e4a-69609bf28a9c service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Neutron deleted interface 7f4f6f32-7ee7-4884-86c4-27ecf7daca85; detaching it from the instance and deleting it from the info cache [ 1175.738655] env[61991]: DEBUG nova.network.neutron [req-a202ab54-b25b-4ed6-bc9f-eee14029d954 req-4a9384f9-a680-440d-8e4a-69609bf28a9c service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.762781] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1176.222751] env[61991]: DEBUG nova.network.neutron [-] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.241343] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecde184f-cbe0-457f-aa4f-5f2f1a80d6c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.250424] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef3d3c9-3563-4908-8aa8-4e7d48f9821f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.268207] env[61991]: INFO nova.virt.block_device [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Booting with volume b63727a3-5dc2-4d03-bee3-469f624cbe25 at /dev/sda [ 1176.273817] env[61991]: DEBUG nova.compute.manager [req-a202ab54-b25b-4ed6-bc9f-eee14029d954 req-4a9384f9-a680-440d-8e4a-69609bf28a9c service nova] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Detach interface failed, port_id=7f4f6f32-7ee7-4884-86c4-27ecf7daca85, reason: Instance 84a0dee7-c33b-494a-ad38-83da6ab44ce2 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1176.303928] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02f3b6aa-9275-4d92-857d-37ef3bd0e095 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.311989] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c79537-5f56-4df6-a9db-4f2de66a57e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.334480] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35b02885-00fd-4f14-bb20-c5b671e0097b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.341460] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea025c1-2dbf-4c0d-bb53-0282c7298941 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.363925] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c860632-c110-4df5-b404-8048ba5d43e7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.369539] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6766b998-c4c6-4bdf-a749-c5a3e0f76d69 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.382204] env[61991]: DEBUG nova.virt.block_device [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updating existing volume attachment record: f1972c54-2a83-4492-bb65-ff5a4f5888df {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1176.450510] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.450714] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquired lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.450894] env[61991]: DEBUG nova.network.neutron [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1176.568252] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 84a0dee7-c33b-494a-ad38-83da6ab44ce2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1176.568415] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance f42c1c55-3635-460e-a106-e63e57426fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1176.568537] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1176.568718] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1176.568859] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1176.614403] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9530030d-7cc2-458c-b5b2-88453060852e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.622332] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99ee0df-4237-49d3-a110-00f61928ded2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.651677] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f8d7ce-5ccf-4fa5-8028-6c96f5afff23 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.658733] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cd70cd-7629-41cc-9dac-bd4c20db20d9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.671316] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.725351] env[61991]: INFO nova.compute.manager [-] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Took 1.22 seconds to deallocate network for instance. [ 1176.973707] env[61991]: DEBUG nova.compute.manager [req-c5d7c465-17e9-4677-a359-ce50befe7c7f req-fef613e2-f080-48d7-9d3b-1aeee83c0636 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Received event network-vif-plugged-e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.974050] env[61991]: DEBUG oslo_concurrency.lockutils [req-c5d7c465-17e9-4677-a359-ce50befe7c7f req-fef613e2-f080-48d7-9d3b-1aeee83c0636 service nova] Acquiring lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.974187] env[61991]: DEBUG oslo_concurrency.lockutils [req-c5d7c465-17e9-4677-a359-ce50befe7c7f req-fef613e2-f080-48d7-9d3b-1aeee83c0636 service nova] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.974368] env[61991]: DEBUG oslo_concurrency.lockutils [req-c5d7c465-17e9-4677-a359-ce50befe7c7f req-fef613e2-f080-48d7-9d3b-1aeee83c0636 service nova] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.974542] env[61991]: DEBUG nova.compute.manager [req-c5d7c465-17e9-4677-a359-ce50befe7c7f req-fef613e2-f080-48d7-9d3b-1aeee83c0636 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] No waiting events found dispatching network-vif-plugged-e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1176.974710] env[61991]: WARNING nova.compute.manager [req-c5d7c465-17e9-4677-a359-ce50befe7c7f req-fef613e2-f080-48d7-9d3b-1aeee83c0636 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Received unexpected event network-vif-plugged-e14b4c9e-77e1-43d7-8a41-a2cf392134a0 for instance with vm_state building and task_state block_device_mapping. [ 1177.058870] env[61991]: DEBUG nova.network.neutron [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Successfully updated port: e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1177.164719] env[61991]: DEBUG nova.network.neutron [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [{"id": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "address": "fa:16:3e:c0:1c:8c", "network": {"id": "5979971a-c809-46a8-8b8b-3a41a2297f91", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1012489079-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "870d2c2c0e554180b190b88bdab5fc2d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3ccbdbb-8b49-4a26-913f-2a448b72280f", "external-id": "nsx-vlan-transportzone-412", "segmentation_id": 412, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5f87d87-c6", "ovs_interfaceid": "b5f87d87-c6b0-45c2-a7d6-92438277b844", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.174204] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1177.231408] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.564915] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.564987] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquired lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.565269] env[61991]: DEBUG nova.network.neutron [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1177.668541] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Releasing lock "refresh_cache-f42c1c55-3635-460e-a106-e63e57426fd2" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.669944] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d910db5-8dac-4656-a218-a0edee801428 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.677146] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Resuming the VM {{(pid=61991) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1177.677391] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b1be370-78d9-44c5-b5a6-00d5748a2808 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.679270] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1177.679457] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.138s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.679693] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.448s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.679969] env[61991]: DEBUG nova.objects.instance [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lazy-loading 'resources' on Instance uuid 84a0dee7-c33b-494a-ad38-83da6ab44ce2 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.686577] env[61991]: DEBUG oslo_vmware.api [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1177.686577] env[61991]: value = "task-1130523" [ 1177.686577] env[61991]: _type = "Task" [ 1177.686577] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.695199] env[61991]: DEBUG oslo_vmware.api [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130523, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.097698] env[61991]: DEBUG nova.network.neutron [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1178.200429] env[61991]: DEBUG oslo_vmware.api [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130523, 'name': PowerOnVM_Task, 'duration_secs': 0.513472} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.200659] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Resumed the VM {{(pid=61991) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1178.200899] env[61991]: DEBUG nova.compute.manager [None req-d01703f3-e696-4750-bad8-923b925b6ec8 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1178.201859] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda0d2f2-bf36-4e44-a7af-5d503a5b04ab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.247791] env[61991]: DEBUG nova.network.neutron [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updating instance_info_cache with network_info: [{"id": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "address": "fa:16:3e:67:a5:53", "network": {"id": "5c08b170-3f4e-448c-a3c0-bb1f0f8a2842", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1830710133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97761803e1374ac2a1d9028c79cbccde", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14b4c9e-77", "ovs_interfaceid": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.253365] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a48335-11dc-4336-81c5-2677f146b20f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.262961] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecf2473-fc26-43c9-9d6e-bd77cd77b90d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.301537] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7df3722-d4b2-483a-9ce2-20d7f930a28f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.312019] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9c321c-58e5-40b0-8547-0241ed48be42 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.330354] env[61991]: DEBUG nova.compute.provider_tree [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.466436] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1178.467260] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1178.467608] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1178.467884] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1178.468219] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1178.468479] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1178.468741] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1178.469111] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1178.469394] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1178.469692] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1178.469988] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1178.470314] env[61991]: DEBUG nova.virt.hardware [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1178.472071] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab8ce5a-799d-4971-bfb9-e069eff61be4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.483869] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de34d767-c1c0-4a2b-8908-a6238e3ceaf4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.750430] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Releasing lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.751500] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance network_info: |[{"id": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "address": "fa:16:3e:67:a5:53", "network": {"id": "5c08b170-3f4e-448c-a3c0-bb1f0f8a2842", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1830710133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97761803e1374ac2a1d9028c79cbccde", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14b4c9e-77", "ovs_interfaceid": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1178.751500] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:a5:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cf5bfbae-a882-4d34-be33-b31e274b3077', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e14b4c9e-77e1-43d7-8a41-a2cf392134a0', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1178.758950] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Creating folder: Project (97761803e1374ac2a1d9028c79cbccde). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1178.759291] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d21332d-e534-4671-97fe-25891ff7df36 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.773289] env[61991]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1178.773501] env[61991]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61991) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1178.773987] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Folder already exists: Project (97761803e1374ac2a1d9028c79cbccde). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1178.774287] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Creating folder: Instances. Parent ref: group-v246984. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1178.774607] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18b65e77-3cfb-4c48-a710-f2239107e561 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.785748] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Created folder: Instances in parent group-v246984. [ 1178.786024] env[61991]: DEBUG oslo.service.loopingcall [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1178.786240] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1178.786457] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0d90886c-1138-406c-a9b2-6c25a5b7e3ae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.806563] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1178.806563] env[61991]: value = "task-1130526" [ 1178.806563] env[61991]: _type = "Task" [ 1178.806563] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.815196] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130526, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.833515] env[61991]: DEBUG nova.scheduler.client.report [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1179.003963] env[61991]: DEBUG nova.compute.manager [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Received event network-changed-e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1179.004289] env[61991]: DEBUG nova.compute.manager [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Refreshing instance network info cache due to event network-changed-e14b4c9e-77e1-43d7-8a41-a2cf392134a0. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1179.004547] env[61991]: DEBUG oslo_concurrency.lockutils [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] Acquiring lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.004713] env[61991]: DEBUG oslo_concurrency.lockutils [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] Acquired lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.004876] env[61991]: DEBUG nova.network.neutron [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Refreshing network info cache for port e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1179.091162] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "f42c1c55-3635-460e-a106-e63e57426fd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.091444] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.091636] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "f42c1c55-3635-460e-a106-e63e57426fd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.091821] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1179.091992] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.096299] env[61991]: INFO nova.compute.manager [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Terminating instance [ 1179.098179] env[61991]: DEBUG nova.compute.manager [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1179.098474] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1179.099274] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bda65e9-8257-4b2a-843b-e2412f7e90dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.106535] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1179.106766] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fb8ae23-3e8c-4357-be8b-e776a0f5bb7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.113540] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1179.113540] env[61991]: value = "task-1130527" [ 1179.113540] env[61991]: _type = "Task" [ 1179.113540] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.121053] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.317014] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130526, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.339249] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.360149] env[61991]: INFO nova.scheduler.client.report [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Deleted allocations for instance 84a0dee7-c33b-494a-ad38-83da6ab44ce2 [ 1179.623604] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.675697] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.676059] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.676229] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1179.676334] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1179.718367] env[61991]: DEBUG nova.network.neutron [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updated VIF entry in instance network info cache for port e14b4c9e-77e1-43d7-8a41-a2cf392134a0. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1179.718713] env[61991]: DEBUG nova.network.neutron [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updating instance_info_cache with network_info: [{"id": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "address": "fa:16:3e:67:a5:53", "network": {"id": "5c08b170-3f4e-448c-a3c0-bb1f0f8a2842", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1830710133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97761803e1374ac2a1d9028c79cbccde", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14b4c9e-77", "ovs_interfaceid": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.816800] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130526, 'name': CreateVM_Task} progress is 99%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.867919] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1a70d538-ddce-4796-a315-b41be84cb125 tempest-ServersTestJSON-1248748401 tempest-ServersTestJSON-1248748401-project-member] Lock "84a0dee7-c33b-494a-ad38-83da6ab44ce2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.480s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.125122] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.180065] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Skipping network cache update for instance because it is being deleted. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1180.180263] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Skipping network cache update for instance because it is Building. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1180.180301] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Didn't find any instances for network info cache update. {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1180.180488] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.180644] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.221787] env[61991]: DEBUG oslo_concurrency.lockutils [req-c8702076-fb3c-4b25-b84d-1325920b172d req-fb908f6f-031b-4f60-a0bf-f4a6b1ca821a service nova] Releasing lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.317916] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130526, 'name': CreateVM_Task, 'duration_secs': 1.363713} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.318110] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1180.318771] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'guest_format': None, 'mount_device': '/dev/sda', 'disk_bus': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246988', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'name': 'volume-b63727a3-5dc2-4d03-bee3-469f624cbe25', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df3b0fba-50ac-4a73-a0e4-f1cfe2705379', 'attached_at': '', 'detached_at': '', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'serial': 'b63727a3-5dc2-4d03-bee3-469f624cbe25'}, 'attachment_id': 'f1972c54-2a83-4492-bb65-ff5a4f5888df', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61991) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1180.318988] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Root volume attach. Driver type: vmdk {{(pid=61991) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1180.319757] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b025539b-759f-4163-b240-661ef250ea71 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.326980] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437e4174-6147-44c6-8e31-46aa9eac1cd5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.333061] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f5969f-7d69-4bef-abef-bf2103082dad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.339115] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-7b8c45ea-b09b-4055-9088-f4ccc66f8ff6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.346198] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1180.346198] env[61991]: value = "task-1130528" [ 1180.346198] env[61991]: _type = "Task" [ 1180.346198] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.353860] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130528, 'name': RelocateVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.625869] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130527, 'name': PowerOffVM_Task, 'duration_secs': 1.150401} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.626373] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1180.626604] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1180.626931] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da9804bb-6001-4909-92cc-689516f90af4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.686451] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1180.686451] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1180.686451] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleting the datastore file [datastore2] f42c1c55-3635-460e-a106-e63e57426fd2 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1180.686569] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2681e49-aaae-4470-b8c8-5ce8b370c7c4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.693814] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for the task: (returnval){ [ 1180.693814] env[61991]: value = "task-1130530" [ 1180.693814] env[61991]: _type = "Task" [ 1180.693814] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.701212] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130530, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.862365] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130528, 'name': RelocateVM_Task, 'duration_secs': 0.352019} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.862768] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1180.863041] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246988', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'name': 'volume-b63727a3-5dc2-4d03-bee3-469f624cbe25', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df3b0fba-50ac-4a73-a0e4-f1cfe2705379', 'attached_at': '', 'detached_at': '', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'serial': 'b63727a3-5dc2-4d03-bee3-469f624cbe25'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1180.863980] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca17261f-958d-4fa0-a3f1-4f2c8a65dca8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.879349] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eea3544-94e0-4ab4-a0c1-6f1fc23ab7c6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.901306] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] volume-b63727a3-5dc2-4d03-bee3-469f624cbe25/volume-b63727a3-5dc2-4d03-bee3-469f624cbe25.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.901622] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-950015c7-c57f-4a9e-a6b6-9b5b3f04c66d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.920948] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1180.920948] env[61991]: value = "task-1130531" [ 1180.920948] env[61991]: _type = "Task" [ 1180.920948] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.930897] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130531, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.204615] env[61991]: DEBUG oslo_vmware.api [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Task: {'id': task-1130530, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149239} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.204940] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1181.205199] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1181.205445] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1181.205681] env[61991]: INFO nova.compute.manager [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Took 2.11 seconds to destroy the instance on the hypervisor. [ 1181.206015] env[61991]: DEBUG oslo.service.loopingcall [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1181.206272] env[61991]: DEBUG nova.compute.manager [-] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1181.206418] env[61991]: DEBUG nova.network.neutron [-] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1181.432389] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130531, 'name': ReconfigVM_Task, 'duration_secs': 0.262993} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.437021] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Reconfigured VM instance instance-00000075 to attach disk [datastore1] volume-b63727a3-5dc2-4d03-bee3-469f624cbe25/volume-b63727a3-5dc2-4d03-bee3-469f624cbe25.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.440818] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d132d3c3-b3de-4e05-a64d-36364d3e51d8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.463951] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1181.463951] env[61991]: value = "task-1130532" [ 1181.463951] env[61991]: _type = "Task" [ 1181.463951] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.473085] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130532, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.860200] env[61991]: DEBUG nova.compute.manager [req-cf6d5520-a366-4422-8338-eefe366022ab req-8fb9d146-5a28-4840-aeae-6a6b6ea0dbff service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Received event network-vif-deleted-b5f87d87-c6b0-45c2-a7d6-92438277b844 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1181.860200] env[61991]: INFO nova.compute.manager [req-cf6d5520-a366-4422-8338-eefe366022ab req-8fb9d146-5a28-4840-aeae-6a6b6ea0dbff service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Neutron deleted interface b5f87d87-c6b0-45c2-a7d6-92438277b844; detaching it from the instance and deleting it from the info cache [ 1181.860200] env[61991]: DEBUG nova.network.neutron [req-cf6d5520-a366-4422-8338-eefe366022ab req-8fb9d146-5a28-4840-aeae-6a6b6ea0dbff service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.973729] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130532, 'name': ReconfigVM_Task, 'duration_secs': 0.112545} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.974054] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246988', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'name': 'volume-b63727a3-5dc2-4d03-bee3-469f624cbe25', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df3b0fba-50ac-4a73-a0e4-f1cfe2705379', 'attached_at': '', 'detached_at': '', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'serial': 'b63727a3-5dc2-4d03-bee3-469f624cbe25'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1181.974593] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e11e7b5b-6e69-4d64-89cb-cf37e091fa53 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.980559] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1181.980559] env[61991]: value = "task-1130533" [ 1181.980559] env[61991]: _type = "Task" [ 1181.980559] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.988315] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130533, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.328316] env[61991]: DEBUG nova.network.neutron [-] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.362086] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1fc7c58-004c-4657-86c6-8e7b799e9622 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.372314] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410bcb13-da4e-4afe-878a-2729e9904724 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.399413] env[61991]: DEBUG nova.compute.manager [req-cf6d5520-a366-4422-8338-eefe366022ab req-8fb9d146-5a28-4840-aeae-6a6b6ea0dbff service nova] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Detach interface failed, port_id=b5f87d87-c6b0-45c2-a7d6-92438277b844, reason: Instance f42c1c55-3635-460e-a106-e63e57426fd2 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1182.490749] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130533, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.831381] env[61991]: INFO nova.compute.manager [-] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Took 1.62 seconds to deallocate network for instance. [ 1182.992499] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130533, 'name': Rename_Task, 'duration_secs': 0.695353} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.992499] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.992499] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f231664-c29a-40a0-b10f-dd73e82a8bcb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.998039] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1182.998039] env[61991]: value = "task-1130534" [ 1182.998039] env[61991]: _type = "Task" [ 1182.998039] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.005369] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.338900] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.339342] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.339648] env[61991]: DEBUG nova.objects.instance [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lazy-loading 'resources' on Instance uuid f42c1c55-3635-460e-a106-e63e57426fd2 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.508493] env[61991]: DEBUG oslo_vmware.api [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130534, 'name': PowerOnVM_Task, 'duration_secs': 0.44171} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.509334] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1183.509334] env[61991]: INFO nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Took 5.04 seconds to spawn the instance on the hypervisor. [ 1183.509334] env[61991]: DEBUG nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1183.509938] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec3ed46-0743-4455-be4d-e2df04ea8a68 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.886759] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0a12b0-b28a-440e-a67f-8bcc78e97387 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.895376] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c2de7b-a3e7-4673-b23f-63d59f3139c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.930159] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86ecef3-10da-464e-a794-7808fe10381f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.938682] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b2b606-f20d-4692-8c3f-f86876de78fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.955306] env[61991]: DEBUG nova.compute.provider_tree [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.026728] env[61991]: INFO nova.compute.manager [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Took 12.39 seconds to build instance. [ 1184.458800] env[61991]: DEBUG nova.scheduler.client.report [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1184.529541] env[61991]: DEBUG oslo_concurrency.lockutils [None req-79b2082f-d1f8-4aa1-adbf-7aea03edc332 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.897s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.964721] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.625s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.995774] env[61991]: INFO nova.scheduler.client.report [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Deleted allocations for instance f42c1c55-3635-460e-a106-e63e57426fd2 [ 1185.469284] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.493849] env[61991]: DEBUG nova.compute.manager [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Received event network-changed-e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1185.493849] env[61991]: DEBUG nova.compute.manager [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Refreshing instance network info cache due to event network-changed-e14b4c9e-77e1-43d7-8a41-a2cf392134a0. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1185.493849] env[61991]: DEBUG oslo_concurrency.lockutils [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] Acquiring lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.493849] env[61991]: DEBUG oslo_concurrency.lockutils [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] Acquired lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.493849] env[61991]: DEBUG nova.network.neutron [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Refreshing network info cache for port e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1185.504101] env[61991]: DEBUG oslo_concurrency.lockutils [None req-1b75eee5-c106-4e0c-aec4-c3e007485355 tempest-ServerActionsTestJSON-904231001 tempest-ServerActionsTestJSON-904231001-project-member] Lock "f42c1c55-3635-460e-a106-e63e57426fd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.412s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.109531] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.109773] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.381939] env[61991]: DEBUG nova.network.neutron [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updated VIF entry in instance network info cache for port e14b4c9e-77e1-43d7-8a41-a2cf392134a0. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1186.382354] env[61991]: DEBUG nova.network.neutron [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updating instance_info_cache with network_info: [{"id": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "address": "fa:16:3e:67:a5:53", "network": {"id": "5c08b170-3f4e-448c-a3c0-bb1f0f8a2842", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1830710133-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.253", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97761803e1374ac2a1d9028c79cbccde", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cf5bfbae-a882-4d34-be33-b31e274b3077", "external-id": "nsx-vlan-transportzone-556", "segmentation_id": 556, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape14b4c9e-77", "ovs_interfaceid": "e14b4c9e-77e1-43d7-8a41-a2cf392134a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.612315] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1186.885511] env[61991]: DEBUG oslo_concurrency.lockutils [req-6d68a0f6-d725-4bfc-b3f6-0039098dee9b req-7381f1c5-9302-49e1-9f41-ca6d39b99a91 service nova] Releasing lock "refresh_cache-df3b0fba-50ac-4a73-a0e4-f1cfe2705379" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.134590] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.134702] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.136664] env[61991]: INFO nova.compute.claims [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1188.186600] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3050ecc1-04be-4fea-9304-ee25bc177303 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.194403] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a962b87e-a028-476f-9cce-ed3fa6fe3a71 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.222878] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87114717-b583-42df-b83b-984e905586da {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.229760] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58acb799-19f4-4f26-b867-cd685ff58f1c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.242562] env[61991]: DEBUG nova.compute.provider_tree [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.745765] env[61991]: DEBUG nova.scheduler.client.report [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1189.251073] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.116s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.251380] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1189.757602] env[61991]: DEBUG nova.compute.utils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1189.759501] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1189.759594] env[61991]: DEBUG nova.network.neutron [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1189.809670] env[61991]: DEBUG nova.policy [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f16b8221426c4d9193b26a665737efdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49893e8554324098b659ce6c24acd5d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1190.089392] env[61991]: DEBUG nova.network.neutron [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Successfully created port: b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1190.263522] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1191.274813] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1191.300402] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.300690] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.300851] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.301056] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.301215] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.301363] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.301569] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.301726] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.301891] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.302066] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.302249] env[61991]: DEBUG nova.virt.hardware [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.303116] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ff3f91-bbe9-4210-87a1-f6874521d7fb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.311371] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8d3063-11f0-42dc-aa0f-d24669389626 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.485278] env[61991]: DEBUG nova.compute.manager [req-5fe3e19e-7877-4410-bb79-90b89973b0b8 req-a4646cad-f516-423d-9152-7f9156d3486b service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Received event network-vif-plugged-b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1191.485492] env[61991]: DEBUG oslo_concurrency.lockutils [req-5fe3e19e-7877-4410-bb79-90b89973b0b8 req-a4646cad-f516-423d-9152-7f9156d3486b service nova] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.485701] env[61991]: DEBUG oslo_concurrency.lockutils [req-5fe3e19e-7877-4410-bb79-90b89973b0b8 req-a4646cad-f516-423d-9152-7f9156d3486b service nova] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.485874] env[61991]: DEBUG oslo_concurrency.lockutils [req-5fe3e19e-7877-4410-bb79-90b89973b0b8 req-a4646cad-f516-423d-9152-7f9156d3486b service nova] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.490013] env[61991]: DEBUG nova.compute.manager [req-5fe3e19e-7877-4410-bb79-90b89973b0b8 req-a4646cad-f516-423d-9152-7f9156d3486b service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] No waiting events found dispatching network-vif-plugged-b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1191.490265] env[61991]: WARNING nova.compute.manager [req-5fe3e19e-7877-4410-bb79-90b89973b0b8 req-a4646cad-f516-423d-9152-7f9156d3486b service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Received unexpected event network-vif-plugged-b73189c5-466b-4390-9763-354c3c026e37 for instance with vm_state building and task_state spawning. [ 1191.566508] env[61991]: DEBUG nova.network.neutron [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Successfully updated port: b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1192.068627] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.068773] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.068931] env[61991]: DEBUG nova.network.neutron [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1192.600831] env[61991]: DEBUG nova.network.neutron [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1192.729506] env[61991]: DEBUG nova.network.neutron [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.232300] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.232637] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Instance network_info: |[{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1193.233101] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:76:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b73189c5-466b-4390-9763-354c3c026e37', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1193.240544] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Creating folder: Project (49893e8554324098b659ce6c24acd5d1). Parent ref: group-v246753. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1193.240822] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-950908b2-10ee-4ee1-83d1-d15b91d6d188 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.251803] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Created folder: Project (49893e8554324098b659ce6c24acd5d1) in parent group-v246753. [ 1193.251980] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Creating folder: Instances. Parent ref: group-v246991. {{(pid=61991) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1193.252209] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86718781-b28c-4f59-99a9-cdbe86636e70 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.260616] env[61991]: INFO nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Created folder: Instances in parent group-v246991. [ 1193.260828] env[61991]: DEBUG oslo.service.loopingcall [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1193.261013] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1193.261247] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c3af352-5dd4-4d45-ada0-fb0bb864ab2b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.279665] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1193.279665] env[61991]: value = "task-1130537" [ 1193.279665] env[61991]: _type = "Task" [ 1193.279665] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.286496] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130537, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.511862] env[61991]: DEBUG nova.compute.manager [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Received event network-changed-b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1193.512139] env[61991]: DEBUG nova.compute.manager [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Refreshing instance network info cache due to event network-changed-b73189c5-466b-4390-9763-354c3c026e37. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1193.512324] env[61991]: DEBUG oslo_concurrency.lockutils [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.512487] env[61991]: DEBUG oslo_concurrency.lockutils [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.512676] env[61991]: DEBUG nova.network.neutron [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Refreshing network info cache for port b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1193.789936] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130537, 'name': CreateVM_Task, 'duration_secs': 0.279743} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.789936] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1193.790606] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.790778] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.791108] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1193.791354] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-464ba229-a183-4f9f-a16d-4218985f323b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.795497] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1193.795497] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205a10a-8a22-b993-32d9-4b162fe13060" [ 1193.795497] env[61991]: _type = "Task" [ 1193.795497] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.802576] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205a10a-8a22-b993-32d9-4b162fe13060, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.191523] env[61991]: DEBUG nova.network.neutron [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updated VIF entry in instance network info cache for port b73189c5-466b-4390-9763-354c3c026e37. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1194.191888] env[61991]: DEBUG nova.network.neutron [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.305645] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]5205a10a-8a22-b993-32d9-4b162fe13060, 'name': SearchDatastore_Task, 'duration_secs': 0.009092} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.305977] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.306240] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1194.306473] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.306623] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.306801] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1194.307068] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a67149b-27fb-44dd-a176-48b96e5859ae {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.314642] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1194.314903] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1194.315628] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b45f5e7-0b69-4514-966e-57a0bf6d4800 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.320395] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1194.320395] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d53e66-c888-ce38-722f-da4e184c76c8" [ 1194.320395] env[61991]: _type = "Task" [ 1194.320395] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.327398] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d53e66-c888-ce38-722f-da4e184c76c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.694999] env[61991]: DEBUG oslo_concurrency.lockutils [req-355b39c8-7d78-4009-a0bf-baeab1d7a34a req-8ccfdb50-8790-40fd-8e68-e010a1e2b586 service nova] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.831328] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52d53e66-c888-ce38-722f-da4e184c76c8, 'name': SearchDatastore_Task, 'duration_secs': 0.008196} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.832181] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3789b004-7211-41a9-adcc-5a8dc846a1f0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.837240] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1194.837240] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522957ee-1ad2-56bd-dc4e-d1952f8c79aa" [ 1194.837240] env[61991]: _type = "Task" [ 1194.837240] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.844886] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522957ee-1ad2-56bd-dc4e-d1952f8c79aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.347781] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]522957ee-1ad2-56bd-dc4e-d1952f8c79aa, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.348102] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "[datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.348369] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 60e3238b-a0ef-4ccf-bce1-633badd10e34/60e3238b-a0ef-4ccf-bce1-633badd10e34.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1195.348627] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f126797-1ccf-433e-b470-58f2bef0015c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.355232] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1195.355232] env[61991]: value = "task-1130538" [ 1195.355232] env[61991]: _type = "Task" [ 1195.355232] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.362501] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.864895] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130538, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.365589] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660765} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.365811] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore1] 60e3238b-a0ef-4ccf-bce1-633badd10e34/60e3238b-a0ef-4ccf-bce1-633badd10e34.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1196.366041] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1196.366292] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff4ef371-736b-42a0-88e2-4ad28b36b961 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.372744] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1196.372744] env[61991]: value = "task-1130539" [ 1196.372744] env[61991]: _type = "Task" [ 1196.372744] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.381696] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.882015] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067946} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.882415] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1196.883047] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6efafc5-1f6b-4d46-8587-58f4e7379f07 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.903842] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 60e3238b-a0ef-4ccf-bce1-633badd10e34/60e3238b-a0ef-4ccf-bce1-633badd10e34.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1196.904070] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fe7be24-41fa-4d6e-b504-5e55c9596a08 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.922492] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1196.922492] env[61991]: value = "task-1130540" [ 1196.922492] env[61991]: _type = "Task" [ 1196.922492] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.929676] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130540, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.432638] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130540, 'name': ReconfigVM_Task, 'duration_secs': 0.27145} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.432920] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 60e3238b-a0ef-4ccf-bce1-633badd10e34/60e3238b-a0ef-4ccf-bce1-633badd10e34.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1197.433562] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0022eddc-dbd8-4b02-8240-8c8da9c5c78e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.439521] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1197.439521] env[61991]: value = "task-1130541" [ 1197.439521] env[61991]: _type = "Task" [ 1197.439521] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.447214] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130541, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.949791] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130541, 'name': Rename_Task} progress is 14%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.450626] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130541, 'name': Rename_Task, 'duration_secs': 0.834019} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.450901] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1198.451166] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d921e419-7907-4900-860e-3196e3eecb66 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.456893] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1198.456893] env[61991]: value = "task-1130542" [ 1198.456893] env[61991]: _type = "Task" [ 1198.456893] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.463979] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130542, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.966090] env[61991]: DEBUG oslo_vmware.api [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130542, 'name': PowerOnVM_Task, 'duration_secs': 0.429317} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.966463] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1198.966513] env[61991]: INFO nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1198.966692] env[61991]: DEBUG nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1198.967445] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d8b318-4708-4ead-863b-919a5921c947 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.485936] env[61991]: INFO nova.compute.manager [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Took 12.37 seconds to build instance. [ 1199.987764] env[61991]: DEBUG oslo_concurrency.lockutils [None req-58958002-4a12-4706-8761-dd1560285637 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.878s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.396237] env[61991]: DEBUG nova.compute.manager [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Received event network-changed-b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1200.396237] env[61991]: DEBUG nova.compute.manager [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Refreshing instance network info cache due to event network-changed-b73189c5-466b-4390-9763-354c3c026e37. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1200.396237] env[61991]: DEBUG oslo_concurrency.lockutils [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.396609] env[61991]: DEBUG oslo_concurrency.lockutils [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.396968] env[61991]: DEBUG nova.network.neutron [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Refreshing network info cache for port b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1201.108385] env[61991]: DEBUG nova.network.neutron [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updated VIF entry in instance network info cache for port b73189c5-466b-4390-9763-354c3c026e37. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1201.108759] env[61991]: DEBUG nova.network.neutron [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.611639] env[61991]: DEBUG oslo_concurrency.lockutils [req-ec8d5aa4-a85e-4848-8a09-1596f3ed43a0 req-5ee7076c-280a-4739-a509-1313390af33e service nova] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.746106] env[61991]: INFO nova.compute.manager [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Rebuilding instance [ 1223.784499] env[61991]: DEBUG nova.compute.manager [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1223.785386] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3dacfc-51ec-44a8-97d3-3255e2eb7af9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.297163] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1224.297491] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5da5e840-f3c3-487c-9e1e-04c123ad3756 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.304426] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1224.304426] env[61991]: value = "task-1130543" [ 1224.304426] env[61991]: _type = "Task" [ 1224.304426] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.313903] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.813952] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130543, 'name': PowerOffVM_Task, 'duration_secs': 0.166668} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.814305] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1224.814900] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1224.815151] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02633ea1-a151-4efc-b7f0-8e397fd91e59 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.820889] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1224.820889] env[61991]: value = "task-1130544" [ 1224.820889] env[61991]: _type = "Task" [ 1224.820889] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.827875] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.330936] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] VM already powered off {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1225.331189] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1225.331387] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246988', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'name': 'volume-b63727a3-5dc2-4d03-bee3-469f624cbe25', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df3b0fba-50ac-4a73-a0e4-f1cfe2705379', 'attached_at': '', 'detached_at': '', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'serial': 'b63727a3-5dc2-4d03-bee3-469f624cbe25'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1225.332146] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ba987c-29b4-4eb8-b696-e0e1bb0936e3 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.357626] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680214a4-f62b-4b7b-93e3-85643b866212 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.363574] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a1ad73-ddce-4d5f-9ba0-8ef6c96bce6d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.382029] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a98233-847d-444c-8466-aa8727b609b2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.396188] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] The volume has not been displaced from its original location: [datastore1] volume-b63727a3-5dc2-4d03-bee3-469f624cbe25/volume-b63727a3-5dc2-4d03-bee3-469f624cbe25.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1225.401299] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1225.401597] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d7d75fd-3a01-4bc0-8006-b6d661fa47ea {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.418993] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1225.418993] env[61991]: value = "task-1130545" [ 1225.418993] env[61991]: _type = "Task" [ 1225.418993] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.429205] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130545, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.928365] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130545, 'name': ReconfigVM_Task, 'duration_secs': 0.180731} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.928685] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1225.933220] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64724db8-fc55-46d0-a955-0bb674308c4d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.947566] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1225.947566] env[61991]: value = "task-1130546" [ 1225.947566] env[61991]: _type = "Task" [ 1225.947566] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.954941] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.458097] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130546, 'name': ReconfigVM_Task, 'duration_secs': 0.097174} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.458447] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246988', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'name': 'volume-b63727a3-5dc2-4d03-bee3-469f624cbe25', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'df3b0fba-50ac-4a73-a0e4-f1cfe2705379', 'attached_at': '', 'detached_at': '', 'volume_id': 'b63727a3-5dc2-4d03-bee3-469f624cbe25', 'serial': 'b63727a3-5dc2-4d03-bee3-469f624cbe25'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1226.458723] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1226.459475] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4350f2f9-6a4a-467d-ae79-6e7ff2ab6245 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.466067] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1226.466295] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6532b093-b525-44bb-84b3-9d6a9dd1ad56 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.522414] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1226.522639] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1226.522858] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Deleting the datastore file [datastore1] df3b0fba-50ac-4a73-a0e4-f1cfe2705379 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1226.523177] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78b03ec2-eab2-44af-8cb7-1e695b61a900 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.529950] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for the task: (returnval){ [ 1226.529950] env[61991]: value = "task-1130548" [ 1226.529950] env[61991]: _type = "Task" [ 1226.529950] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.538199] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130548, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.040503] env[61991]: DEBUG oslo_vmware.api [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Task: {'id': task-1130548, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.073563} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.040826] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1227.040922] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1227.041130] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1227.096797] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1227.097133] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1baa707d-583e-43ec-884c-16453006f24b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.106136] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d495b16-b4be-467e-a38e-fe0b35c4b2dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.129540] env[61991]: ERROR nova.compute.manager [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Failed to detach volume b63727a3-5dc2-4d03-bee3-469f624cbe25 from /dev/sda: nova.exception.InstanceNotFound: Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 could not be found. [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Traceback (most recent call last): [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self.driver.rebuild(**kwargs) [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] raise NotImplementedError() [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] NotImplementedError [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] During handling of the above exception, another exception occurred: [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Traceback (most recent call last): [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self.driver.detach_volume(context, old_connection_info, [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] return self._volumeops.detach_volume(connection_info, instance) [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._detach_volume_vmdk(connection_info, instance) [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] stable_ref.fetch_moref(session) [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] nova.exception.InstanceNotFound: Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 could not be found. [ 1227.129540] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.247610] env[61991]: DEBUG nova.compute.utils [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Build of instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 aborted: Failed to rebuild volume backed instance. {{(pid=61991) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1227.249888] env[61991]: ERROR nova.compute.manager [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 aborted: Failed to rebuild volume backed instance. [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Traceback (most recent call last): [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self.driver.rebuild(**kwargs) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] raise NotImplementedError() [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] NotImplementedError [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] During handling of the above exception, another exception occurred: [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Traceback (most recent call last): [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._detach_root_volume(context, instance, root_bdm) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] with excutils.save_and_reraise_exception(): [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self.force_reraise() [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] raise self.value [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self.driver.detach_volume(context, old_connection_info, [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] return self._volumeops.detach_volume(connection_info, instance) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._detach_volume_vmdk(connection_info, instance) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] stable_ref.fetch_moref(session) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] nova.exception.InstanceNotFound: Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 could not be found. [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] During handling of the above exception, another exception occurred: [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Traceback (most recent call last): [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] yield [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1227.249888] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._do_rebuild_instance_with_claim( [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._do_rebuild_instance( [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._rebuild_default_impl(**kwargs) [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] self._rebuild_volume_backed_instance( [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] raise exception.BuildAbortException( [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] nova.exception.BuildAbortException: Build of instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 aborted: Failed to rebuild volume backed instance. [ 1227.251168] env[61991]: ERROR nova.compute.manager [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] [ 1229.264153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.264475] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.292704] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8800e641-dd1e-4bb9-905d-c879ba559f7c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.300209] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac83e5ab-56dd-4656-a670-7d6c2cc4af3d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.331989] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8fd363-567d-4aa2-ab1c-e626db4b132b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.342634] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9447b8d9-ddee-4217-a83b-33a4717c9b45 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.356734] env[61991]: DEBUG nova.compute.provider_tree [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1229.837918] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.838159] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.838428] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.838614] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.838794] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.841158] env[61991]: INFO nova.compute.manager [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Terminating instance [ 1229.843195] env[61991]: DEBUG nova.compute.manager [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1229.843472] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e10d0f2c-8cfe-4056-bcc4-a1a6f03fc3be {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.854098] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e935c1-2bf2-4add-9df7-59f3a657129b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.864869] env[61991]: DEBUG nova.scheduler.client.report [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1229.880903] env[61991]: WARNING nova.virt.vmwareapi.driver [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 could not be found. [ 1229.881104] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1229.881887] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-164c7da2-893c-4780-a750-1937629e6408 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.889835] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73689a0-1dc2-42a6-b808-2c031898baad {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.912351] env[61991]: WARNING nova.virt.vmwareapi.vmops [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 could not be found. [ 1229.912536] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1229.912715] env[61991]: INFO nova.compute.manager [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Took 0.07 seconds to destroy the instance on the hypervisor. [ 1229.912947] env[61991]: DEBUG oslo.service.loopingcall [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1229.913169] env[61991]: DEBUG nova.compute.manager [-] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1229.913279] env[61991]: DEBUG nova.network.neutron [-] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1230.353894] env[61991]: DEBUG nova.compute.manager [req-4ea5f666-5d0d-406f-8ca0-27d9f6af7f1d req-a4e87e00-865c-47bc-9276-6eac8a5699cb service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Received event network-vif-deleted-e14b4c9e-77e1-43d7-8a41-a2cf392134a0 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1230.354176] env[61991]: INFO nova.compute.manager [req-4ea5f666-5d0d-406f-8ca0-27d9f6af7f1d req-a4e87e00-865c-47bc-9276-6eac8a5699cb service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Neutron deleted interface e14b4c9e-77e1-43d7-8a41-a2cf392134a0; detaching it from the instance and deleting it from the info cache [ 1230.354294] env[61991]: DEBUG nova.network.neutron [req-4ea5f666-5d0d-406f-8ca0-27d9f6af7f1d req-a4e87e00-865c-47bc-9276-6eac8a5699cb service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.369054] env[61991]: DEBUG oslo_concurrency.lockutils [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.104s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.369199] env[61991]: INFO nova.compute.manager [None req-257dd28a-7346-4027-a431-2d892e548261 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Successfully reverted task state from rebuilding on failure for instance. [ 1230.833365] env[61991]: DEBUG nova.network.neutron [-] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.856980] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb925df8-4e18-4595-9711-6885378c6d78 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.867022] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006cc501-4877-4a11-b6ef-d2409007f2f4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.892756] env[61991]: DEBUG nova.compute.manager [req-4ea5f666-5d0d-406f-8ca0-27d9f6af7f1d req-a4e87e00-865c-47bc-9276-6eac8a5699cb service nova] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Detach interface failed, port_id=e14b4c9e-77e1-43d7-8a41-a2cf392134a0, reason: Instance df3b0fba-50ac-4a73-a0e4-f1cfe2705379 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1231.337010] env[61991]: INFO nova.compute.manager [-] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Took 1.42 seconds to deallocate network for instance. [ 1231.884331] env[61991]: INFO nova.compute.manager [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Took 0.55 seconds to detach 1 volumes for instance. [ 1231.886462] env[61991]: DEBUG nova.compute.manager [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Deleting volume: b63727a3-5dc2-4d03-bee3-469f624cbe25 {{(pid=61991) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1232.431687] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.431994] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.432247] env[61991]: DEBUG nova.objects.instance [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lazy-loading 'resources' on Instance uuid df3b0fba-50ac-4a73-a0e4-f1cfe2705379 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.967439] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d711285c-1b93-421f-a1f4-5709e9c24a0e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.970561] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1232.975549] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05b7b0f-e8b2-419c-aa3a-813c3cdabaa0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.004153] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edda549-e6aa-470c-a236-ea0ac3a30420 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.010680] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115e5282-0e3f-473e-82dd-f2f7421dff38 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.022911] env[61991]: DEBUG nova.compute.provider_tree [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.526340] env[61991]: DEBUG nova.scheduler.client.report [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1233.969619] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1233.969902] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1234.031014] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.548293] env[61991]: DEBUG oslo_concurrency.lockutils [None req-94abc6e5-07b3-4aa3-8bf9-d4ecb80e5172 tempest-ServerActionsV293TestJSON-300973051 tempest-ServerActionsV293TestJSON-300973051-project-member] Lock "df3b0fba-50ac-4a73-a0e4-f1cfe2705379" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.710s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.969766] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1235.970239] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1236.965232] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1236.968828] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1236.968974] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1236.969105] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1237.512180] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1237.512442] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.512683] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1237.512933] env[61991]: DEBUG nova.objects.instance [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lazy-loading 'info_cache' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.286317] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.286562] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.790063] env[61991]: DEBUG nova.compute.utils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1239.251099] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.292891] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.753739] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1239.753953] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1239.754194] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1239.754363] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1239.754522] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.258243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.258600] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.258732] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.258876] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1240.259819] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07dc6b0e-1d39-45d5-afcf-226dcffd8ded {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.268099] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a0e224-2204-4ed5-86f6-7639e8344631 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.282326] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efce9cf-b341-4b9f-93b2-ef803451ff3a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.288507] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b45298-6529-43a6-a7ca-31f44f7fc40f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.316221] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181297MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1240.316366] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.316582] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.361413] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.361670] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.361905] env[61991]: INFO nova.compute.manager [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Attaching volume 2fea34d8-dc05-4952-b5bb-dead6ce77b89 to /dev/sdb [ 1240.391929] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c4c2f0-65b3-4ff0-904f-5c4070817b24 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.399144] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d6213d-2d1c-49f2-a3fd-a17ab4711ad1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.412383] env[61991]: DEBUG nova.virt.block_device [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating existing volume attachment record: 3d7bc960-6685-4c62-957a-6bc06702ea56 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1241.341358] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 60e3238b-a0ef-4ccf-bce1-633badd10e34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1241.341596] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1241.341706] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1241.366336] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bd9b48-a4f7-4fd1-9e8a-8245c9a281d7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.373779] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59750c17-973d-41e4-bd27-cbb5f7fc114c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.403764] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb8d73c-3c58-4b22-bce3-088cd7aadc57 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.410633] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4339ad71-bdd1-424b-adaa-f403e70134b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.423012] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1241.926303] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1242.431779] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1242.432243] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.991170] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1244.991170] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246995', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'name': 'volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60e3238b-a0ef-4ccf-bce1-633badd10e34', 'attached_at': '', 'detached_at': '', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'serial': '2fea34d8-dc05-4952-b5bb-dead6ce77b89'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1244.991170] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b99c78-7e6b-4247-9200-dd70b584a555 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.991170] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86a83e3-507e-4ba2-b2ea-525a7430ce32 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.014977] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89/volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1245.015331] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5872b502-c1d6-4f7e-9127-ff1df8c788fd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.033977] env[61991]: DEBUG oslo_vmware.api [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1245.033977] env[61991]: value = "task-1130554" [ 1245.033977] env[61991]: _type = "Task" [ 1245.033977] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.041317] env[61991]: DEBUG oslo_vmware.api [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130554, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.543174] env[61991]: DEBUG oslo_vmware.api [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130554, 'name': ReconfigVM_Task, 'duration_secs': 0.327553} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.543465] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Reconfigured VM instance instance-00000076 to attach disk [datastore2] volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89/volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1245.548222] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e385c0a3-01aa-41e3-a16b-76126a946090 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.562348] env[61991]: DEBUG oslo_vmware.api [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1245.562348] env[61991]: value = "task-1130555" [ 1245.562348] env[61991]: _type = "Task" [ 1245.562348] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.571770] env[61991]: DEBUG oslo_vmware.api [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130555, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.072136] env[61991]: DEBUG oslo_vmware.api [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130555, 'name': ReconfigVM_Task, 'duration_secs': 0.13036} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.072461] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246995', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'name': 'volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60e3238b-a0ef-4ccf-bce1-633badd10e34', 'attached_at': '', 'detached_at': '', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'serial': '2fea34d8-dc05-4952-b5bb-dead6ce77b89'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1247.108868] env[61991]: DEBUG nova.objects.instance [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.615392] env[61991]: DEBUG oslo_concurrency.lockutils [None req-c29cdc59-1aec-43e6-ae80-73faa7ea44ab tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.254s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.684174] env[61991]: DEBUG oslo_concurrency.lockutils [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.684424] env[61991]: DEBUG oslo_concurrency.lockutils [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.684628] env[61991]: DEBUG nova.compute.manager [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1247.685519] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7734baf4-a5b2-40c2-a070-4670b2e931d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.692661] env[61991]: DEBUG nova.compute.manager [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1247.693228] env[61991]: DEBUG nova.objects.instance [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1248.198296] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1248.198703] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0034de5-b1a6-4d70-9862-f6cd71b3a422 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.205875] env[61991]: DEBUG oslo_vmware.api [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1248.205875] env[61991]: value = "task-1130556" [ 1248.205875] env[61991]: _type = "Task" [ 1248.205875] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.213958] env[61991]: DEBUG oslo_vmware.api [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130556, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.716484] env[61991]: DEBUG oslo_vmware.api [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130556, 'name': PowerOffVM_Task, 'duration_secs': 0.27353} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.716709] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1248.716896] env[61991]: DEBUG nova.compute.manager [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1248.717647] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c56bd03-6710-40c8-a2e4-1099306f0885 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.229262] env[61991]: DEBUG oslo_concurrency.lockutils [None req-30f62ace-d4d9-49df-b479-c0b07c488d70 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.119513] env[61991]: DEBUG nova.objects.instance [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1250.625251] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.625645] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.625645] env[61991]: DEBUG nova.network.neutron [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1250.625788] env[61991]: DEBUG nova.objects.instance [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'info_cache' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1251.129540] env[61991]: DEBUG nova.objects.base [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Object Instance<60e3238b-a0ef-4ccf-bce1-633badd10e34> lazy-loaded attributes: flavor,info_cache {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1251.833739] env[61991]: DEBUG nova.network.neutron [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.336473] env[61991]: DEBUG oslo_concurrency.lockutils [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1252.840771] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1252.841155] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ff162e2-e714-42c1-a5cc-0816df40aa6a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.849761] env[61991]: DEBUG oslo_vmware.api [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1252.849761] env[61991]: value = "task-1130557" [ 1252.849761] env[61991]: _type = "Task" [ 1252.849761] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.858279] env[61991]: DEBUG oslo_vmware.api [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130557, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.360122] env[61991]: DEBUG oslo_vmware.api [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130557, 'name': PowerOnVM_Task, 'duration_secs': 0.366035} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.360426] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1253.360607] env[61991]: DEBUG nova.compute.manager [None req-4744e53c-a71c-4d9a-af42-90b3a8671bea tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1253.361390] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712cf942-f331-4caa-9954-cd72f5058a98 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.203751] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1290.204181] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.707554] env[61991]: INFO nova.compute.manager [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Detaching volume 2fea34d8-dc05-4952-b5bb-dead6ce77b89 [ 1290.736175] env[61991]: INFO nova.virt.block_device [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Attempting to driver detach volume 2fea34d8-dc05-4952-b5bb-dead6ce77b89 from mountpoint /dev/sdb [ 1290.736425] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1290.736615] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246995', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'name': 'volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60e3238b-a0ef-4ccf-bce1-633badd10e34', 'attached_at': '', 'detached_at': '', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'serial': '2fea34d8-dc05-4952-b5bb-dead6ce77b89'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1290.737504] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4360a97a-0fc1-4aca-81a3-c2c9537c46f9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.758525] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cc6dfd-e6e1-4457-b237-ddba4ab881b0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.764767] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb6367e-484a-49e0-98cf-faf9dd2e6b8c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.784093] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be828104-7b90-4722-871f-4732f7d99c67 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.797433] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] The volume has not been displaced from its original location: [datastore2] volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89/volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1290.802445] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Reconfiguring VM instance instance-00000076 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1290.802681] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b6e2189-e7f8-4ae2-9271-28536f4c6e5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.819556] env[61991]: DEBUG oslo_vmware.api [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1290.819556] env[61991]: value = "task-1130558" [ 1290.819556] env[61991]: _type = "Task" [ 1290.819556] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.826856] env[61991]: DEBUG oslo_vmware.api [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.328609] env[61991]: DEBUG oslo_vmware.api [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130558, 'name': ReconfigVM_Task, 'duration_secs': 0.211384} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.328929] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Reconfigured VM instance instance-00000076 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1291.333452] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccefc593-2c78-446a-8459-62c1b6dad7ce {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.348648] env[61991]: DEBUG oslo_vmware.api [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1291.348648] env[61991]: value = "task-1130559" [ 1291.348648] env[61991]: _type = "Task" [ 1291.348648] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.356325] env[61991]: DEBUG oslo_vmware.api [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130559, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.859785] env[61991]: DEBUG oslo_vmware.api [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130559, 'name': ReconfigVM_Task, 'duration_secs': 0.122775} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.860106] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246995', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'name': 'volume-2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '60e3238b-a0ef-4ccf-bce1-633badd10e34', 'attached_at': '', 'detached_at': '', 'volume_id': '2fea34d8-dc05-4952-b5bb-dead6ce77b89', 'serial': '2fea34d8-dc05-4952-b5bb-dead6ce77b89'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1292.400602] env[61991]: DEBUG nova.objects.instance [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1293.370160] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1293.408284] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b0b312c9-4c9a-4055-87de-5cdbb21ec955 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.204s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1293.409359] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.039s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1293.409546] env[61991]: DEBUG nova.compute.manager [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1293.410583] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36f287d-15df-4e02-b788-0b92c7251ea7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.417336] env[61991]: DEBUG nova.compute.manager [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61991) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1293.417882] env[61991]: DEBUG nova.objects.instance [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1293.922505] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1293.922771] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50033d1b-4f83-46cd-a529-567285f10d7d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.930298] env[61991]: DEBUG oslo_vmware.api [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1293.930298] env[61991]: value = "task-1130560" [ 1293.930298] env[61991]: _type = "Task" [ 1293.930298] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.938608] env[61991]: DEBUG oslo_vmware.api [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.439507] env[61991]: DEBUG oslo_vmware.api [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130560, 'name': PowerOffVM_Task, 'duration_secs': 0.182507} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.439884] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1294.439937] env[61991]: DEBUG nova.compute.manager [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1294.440672] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54e4e58-00c3-49aa-949f-6e16787da92a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.951862] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e88b1058-0d3e-480c-aba1-645296e7e9d6 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.770066] env[61991]: DEBUG nova.objects.instance [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1296.275857] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1296.276056] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1296.276228] env[61991]: DEBUG nova.network.neutron [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1296.276408] env[61991]: DEBUG nova.objects.instance [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'info_cache' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1296.779433] env[61991]: DEBUG nova.objects.base [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Object Instance<60e3238b-a0ef-4ccf-bce1-633badd10e34> lazy-loaded attributes: flavor,info_cache {{(pid=61991) wrapper /opt/stack/nova/nova/objects/base.py:145}} [ 1297.486499] env[61991]: DEBUG nova.network.neutron [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1297.647361] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1297.647559] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1297.647710] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1297.647863] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1297.648013] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1297.965411] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1297.989344] env[61991]: DEBUG oslo_concurrency.lockutils [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1298.492687] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1298.492952] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bf203b3-cc11-41d1-a706-32be42ea1e36 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.500655] env[61991]: DEBUG oslo_vmware.api [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1298.500655] env[61991]: value = "task-1130561" [ 1298.500655] env[61991]: _type = "Task" [ 1298.500655] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.509112] env[61991]: DEBUG oslo_vmware.api [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.969586] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1298.970036] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1298.970036] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1299.010395] env[61991]: DEBUG oslo_vmware.api [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130561, 'name': PowerOnVM_Task, 'duration_secs': 0.431732} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.010725] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1299.010914] env[61991]: DEBUG nova.compute.manager [None req-e512ddee-4c92-4769-8293-72778f96f09f tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1299.011699] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c5b977-8544-4b4d-b090-d9f3cfbe5ce9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.473463] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1299.473663] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.473771] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1299.473964] env[61991]: DEBUG nova.objects.instance [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lazy-loading 'info_cache' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1301.187038] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [{"id": "b73189c5-466b-4390-9763-354c3c026e37", "address": "fa:16:3e:83:76:86", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb73189c5-46", "ovs_interfaceid": "b73189c5-466b-4390-9763-354c3c026e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1301.690392] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-60e3238b-a0ef-4ccf-bce1-633badd10e34" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1301.690610] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1301.690825] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1301.690988] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1301.691170] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1302.194810] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.195218] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1302.195218] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1302.195390] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1302.196286] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33f993b-0d77-4a93-8804-a5f3ab9db522 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.205809] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d0956e-8a6a-4168-9c67-e2db6ff9b1a1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.219545] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd0943c-8a71-47a2-9056-1acbc245ddab {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.225488] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a53b26-2187-4527-8500-4fde4e814094 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.252766] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181458MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1302.252926] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1302.253108] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1303.278674] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 60e3238b-a0ef-4ccf-bce1-633badd10e34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1303.279047] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1303.279092] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1303.305119] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc62384-dd69-4972-ab2d-47625977df79 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.312755] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880fb33f-9b8c-4f23-b178-227a100762cd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.342292] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c944e5cf-1bc0-4dcb-b0b7-11e7a28281e8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.349021] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de87dcde-5c07-41f9-bedb-dcad924e66ba {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.361366] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1303.864059] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1303.865337] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1303.865513] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.612s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.861236] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.845620] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.846054] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.846103] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "60e3238b-a0ef-4ccf-bce1-633badd10e34-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.846327] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.846509] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.849063] env[61991]: INFO nova.compute.manager [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Terminating instance [ 1330.850872] env[61991]: DEBUG nova.compute.manager [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1330.851086] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1330.851934] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8e3551-9435-4c18-bb68-31b608535043 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.859370] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1330.859842] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9bd603a-0c08-4c3c-82c5-dc0056eaea9c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.865771] env[61991]: DEBUG oslo_vmware.api [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1330.865771] env[61991]: value = "task-1130562" [ 1330.865771] env[61991]: _type = "Task" [ 1330.865771] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.873816] env[61991]: DEBUG oslo_vmware.api [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.375881] env[61991]: DEBUG oslo_vmware.api [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130562, 'name': PowerOffVM_Task, 'duration_secs': 0.176815} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.376173] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1331.376420] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1331.376672] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5575cd94-a9e6-4020-b210-396c8c62e587 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.437152] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1331.437485] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Deleting contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1331.437762] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Deleting the datastore file [datastore1] 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1331.438097] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c33724bb-f7d5-428c-a68b-f8c65e118459 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.444379] env[61991]: DEBUG oslo_vmware.api [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1331.444379] env[61991]: value = "task-1130564" [ 1331.444379] env[61991]: _type = "Task" [ 1331.444379] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.451680] env[61991]: DEBUG oslo_vmware.api [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130564, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.954584] env[61991]: DEBUG oslo_vmware.api [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130564, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147117} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.954972] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1331.955039] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Deleted contents of the VM from datastore datastore1 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1331.955209] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1331.955456] env[61991]: INFO nova.compute.manager [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1331.955741] env[61991]: DEBUG oslo.service.loopingcall [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1331.955939] env[61991]: DEBUG nova.compute.manager [-] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1331.956048] env[61991]: DEBUG nova.network.neutron [-] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1332.373160] env[61991]: DEBUG nova.compute.manager [req-9456cf71-08d5-4366-ac6e-0ee3a3154dd3 req-ec44064d-d373-484b-be0f-a14b0a7332a7 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Received event network-vif-deleted-b73189c5-466b-4390-9763-354c3c026e37 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1332.373350] env[61991]: INFO nova.compute.manager [req-9456cf71-08d5-4366-ac6e-0ee3a3154dd3 req-ec44064d-d373-484b-be0f-a14b0a7332a7 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Neutron deleted interface b73189c5-466b-4390-9763-354c3c026e37; detaching it from the instance and deleting it from the info cache [ 1332.373532] env[61991]: DEBUG nova.network.neutron [req-9456cf71-08d5-4366-ac6e-0ee3a3154dd3 req-ec44064d-d373-484b-be0f-a14b0a7332a7 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.849793] env[61991]: DEBUG nova.network.neutron [-] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1332.875485] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3426a6ca-180d-406d-8b13-3ae608864545 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.885133] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abca84e4-b088-4eed-a53d-fa85bb3ff95a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.906834] env[61991]: DEBUG nova.compute.manager [req-9456cf71-08d5-4366-ac6e-0ee3a3154dd3 req-ec44064d-d373-484b-be0f-a14b0a7332a7 service nova] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Detach interface failed, port_id=b73189c5-466b-4390-9763-354c3c026e37, reason: Instance 60e3238b-a0ef-4ccf-bce1-633badd10e34 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1333.353327] env[61991]: INFO nova.compute.manager [-] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Took 1.40 seconds to deallocate network for instance. [ 1333.860656] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1333.860979] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1333.861233] env[61991]: DEBUG nova.objects.instance [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'resources' on Instance uuid 60e3238b-a0ef-4ccf-bce1-633badd10e34 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1334.383629] env[61991]: DEBUG nova.scheduler.client.report [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Refreshing inventories for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1334.398775] env[61991]: DEBUG nova.scheduler.client.report [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Updating ProviderTree inventory for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1334.399037] env[61991]: DEBUG nova.compute.provider_tree [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Updating inventory in ProviderTree for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1334.409301] env[61991]: DEBUG nova.scheduler.client.report [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Refreshing aggregate associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, aggregates: None {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1334.425907] env[61991]: DEBUG nova.scheduler.client.report [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Refreshing trait associations for resource provider d748992a-e0bf-4ec2-9c17-0e373360e5a3, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61991) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1334.451318] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9bfe4f-9999-4fc7-aa80-dab1b430f474 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.458853] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd568cc2-890c-4cf9-aeeb-d1508e4bb919 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.489187] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4251c921-378a-4929-9300-ffa62085fddd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.495916] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c2cc1a-fd7d-4e10-a4c0-7b070f555e8a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.508736] env[61991]: DEBUG nova.compute.provider_tree [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1335.012377] env[61991]: DEBUG nova.scheduler.client.report [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1335.517992] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1335.537838] env[61991]: INFO nova.scheduler.client.report [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Deleted allocations for instance 60e3238b-a0ef-4ccf-bce1-633badd10e34 [ 1336.046376] env[61991]: DEBUG oslo_concurrency.lockutils [None req-52d612e2-16ba-4eb3-af2f-b279584b9ea5 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "60e3238b-a0ef-4ccf-bce1-633badd10e34" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.200s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1337.749946] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1337.749946] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.252275] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Starting instance... {{(pid=61991) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1338.773153] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.773419] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1338.774928] env[61991]: INFO nova.compute.claims [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1339.809391] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b1031e-5c7a-49a0-8192-72ff5b71145a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.817060] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da729164-5eb8-46f1-9b40-a7f1aa4deacb {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.845850] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e96d96-92d4-4751-8251-53aaea584b8b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.852497] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6feb22ed-ac0d-4b37-ad0b-7100e332908d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.864867] env[61991]: DEBUG nova.compute.provider_tree [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.368183] env[61991]: DEBUG nova.scheduler.client.report [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1340.873500] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.100s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.874020] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Start building networks asynchronously for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1341.380396] env[61991]: DEBUG nova.compute.utils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1341.382510] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Allocating IP information in the background. {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1341.382715] env[61991]: DEBUG nova.network.neutron [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] allocate_for_instance() {{(pid=61991) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1341.436340] env[61991]: DEBUG nova.policy [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f16b8221426c4d9193b26a665737efdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49893e8554324098b659ce6c24acd5d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61991) authorize /opt/stack/nova/nova/policy.py:201}} [ 1341.718482] env[61991]: DEBUG nova.network.neutron [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Successfully created port: 2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1341.886475] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Start building block device mappings for instance. {{(pid=61991) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1342.896127] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Start spawning the instance on the hypervisor. {{(pid=61991) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1342.920569] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-04T09:15:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-04T09:15:10Z,direct_url=,disk_format='vmdk',id=254d700f-2f5a-49a3-8762-cec07162124a,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='351f52655e4f4262b51dd57762a56c2e',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-04T09:15:11Z,virtual_size=,visibility=), allow threads: False {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1342.920815] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Flavor limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1342.920972] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Image limits 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1342.921177] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Flavor pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1342.921326] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Image pref 0:0:0 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1342.921477] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61991) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1342.921687] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1342.921846] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1342.922019] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Got 1 possible topologies {{(pid=61991) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1342.922189] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1342.922385] env[61991]: DEBUG nova.virt.hardware [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61991) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1342.923257] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a0f12e-4519-4347-9a56-34a630d2c242 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.932153] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb0b8c9-4026-41c1-a0ed-db27930f52ec {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.080893] env[61991]: DEBUG nova.compute.manager [req-b5a236ca-ac63-4fce-bc5f-f29a738e3492 req-f8392a27-d0d3-4ef3-bb60-72d4aec2bf39 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Received event network-vif-plugged-2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1343.081139] env[61991]: DEBUG oslo_concurrency.lockutils [req-b5a236ca-ac63-4fce-bc5f-f29a738e3492 req-f8392a27-d0d3-4ef3-bb60-72d4aec2bf39 service nova] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1343.081403] env[61991]: DEBUG oslo_concurrency.lockutils [req-b5a236ca-ac63-4fce-bc5f-f29a738e3492 req-f8392a27-d0d3-4ef3-bb60-72d4aec2bf39 service nova] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1343.081515] env[61991]: DEBUG oslo_concurrency.lockutils [req-b5a236ca-ac63-4fce-bc5f-f29a738e3492 req-f8392a27-d0d3-4ef3-bb60-72d4aec2bf39 service nova] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1343.081678] env[61991]: DEBUG nova.compute.manager [req-b5a236ca-ac63-4fce-bc5f-f29a738e3492 req-f8392a27-d0d3-4ef3-bb60-72d4aec2bf39 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] No waiting events found dispatching network-vif-plugged-2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1343.081827] env[61991]: WARNING nova.compute.manager [req-b5a236ca-ac63-4fce-bc5f-f29a738e3492 req-f8392a27-d0d3-4ef3-bb60-72d4aec2bf39 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Received unexpected event network-vif-plugged-2cfbc2b4-6db8-4951-aae1-61fd8804b599 for instance with vm_state building and task_state spawning. [ 1343.163291] env[61991]: DEBUG nova.network.neutron [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Successfully updated port: 2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1343.664968] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1343.665142] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1343.665293] env[61991]: DEBUG nova.network.neutron [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Building network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1344.196083] env[61991]: DEBUG nova.network.neutron [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1344.313465] env[61991]: DEBUG nova.network.neutron [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [{"id": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "address": "fa:16:3e:5d:ab:b9", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cfbc2b4-6d", "ovs_interfaceid": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.815816] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1344.816155] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Instance network_info: |[{"id": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "address": "fa:16:3e:5d:ab:b9", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cfbc2b4-6d", "ovs_interfaceid": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61991) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1344.816603] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:ab:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cfbc2b4-6db8-4951-aae1-61fd8804b599', 'vif_model': 'vmxnet3'}] {{(pid=61991) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1344.824058] env[61991]: DEBUG oslo.service.loopingcall [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1344.824283] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Creating VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1344.824869] env[61991]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-969bcab3-eec0-40b1-b309-ba1fbd88e903 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.844142] env[61991]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1344.844142] env[61991]: value = "task-1130565" [ 1344.844142] env[61991]: _type = "Task" [ 1344.844142] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.851385] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130565, 'name': CreateVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.106767] env[61991]: DEBUG nova.compute.manager [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Received event network-changed-2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1345.106977] env[61991]: DEBUG nova.compute.manager [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Refreshing instance network info cache due to event network-changed-2cfbc2b4-6db8-4951-aae1-61fd8804b599. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1345.107186] env[61991]: DEBUG oslo_concurrency.lockutils [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] Acquiring lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.107372] env[61991]: DEBUG oslo_concurrency.lockutils [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] Acquired lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.107561] env[61991]: DEBUG nova.network.neutron [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Refreshing network info cache for port 2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1345.353871] env[61991]: DEBUG oslo_vmware.api [-] Task: {'id': task-1130565, 'name': CreateVM_Task, 'duration_secs': 0.292027} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.354144] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Created VM on the ESX host {{(pid=61991) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1345.361104] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.361285] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.361609] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1345.361839] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1026ee2-78b4-4f95-ba16-1ca2af34f488 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.366117] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1345.366117] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52938757-b32d-920b-9832-33ba06cd6d50" [ 1345.366117] env[61991]: _type = "Task" [ 1345.366117] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.374406] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52938757-b32d-920b-9832-33ba06cd6d50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.790061] env[61991]: DEBUG nova.network.neutron [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updated VIF entry in instance network info cache for port 2cfbc2b4-6db8-4951-aae1-61fd8804b599. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1345.790459] env[61991]: DEBUG nova.network.neutron [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [{"id": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "address": "fa:16:3e:5d:ab:b9", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cfbc2b4-6d", "ovs_interfaceid": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1345.876249] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52938757-b32d-920b-9832-33ba06cd6d50, 'name': SearchDatastore_Task, 'duration_secs': 0.009866} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.876533] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1345.876764] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Processing image 254d700f-2f5a-49a3-8762-cec07162124a {{(pid=61991) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1345.876988] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1345.877161] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquired lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1345.877362] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1345.877609] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ec3aa20-cdc7-4099-b790-9a458a5f97b6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.885010] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61991) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1345.885187] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61991) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1345.885818] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f372781e-3939-4e53-a78f-2702b00e8ff6 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.890421] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1345.890421] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db2e9c-5110-83e0-1d8f-71b91f1b7834" [ 1345.890421] env[61991]: _type = "Task" [ 1345.890421] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.896982] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db2e9c-5110-83e0-1d8f-71b91f1b7834, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.293575] env[61991]: DEBUG oslo_concurrency.lockutils [req-07dec8e2-1f7b-48af-b634-54d7b8ace9e7 req-50a8c75a-237f-49cf-bec7-17044c91a9be service nova] Releasing lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1346.400988] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52db2e9c-5110-83e0-1d8f-71b91f1b7834, 'name': SearchDatastore_Task, 'duration_secs': 0.008332} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.401750] env[61991]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da84e280-6613-4680-be4f-d1794f0bf00d {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.406482] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1346.406482] env[61991]: value = "session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c07b1c-959e-9d91-6366-16f5690fdd86" [ 1346.406482] env[61991]: _type = "Task" [ 1346.406482] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.413529] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c07b1c-959e-9d91-6366-16f5690fdd86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.916846] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': session[52341913-c3c7-25c6-a9d2-080f4e5a5690]52c07b1c-959e-9d91-6366-16f5690fdd86, 'name': SearchDatastore_Task, 'duration_secs': 0.00913} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.917085] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Releasing lock "[datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1346.917377] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 9b378b6f-03a3-45a5-a5cc-66e1c6177e58/9b378b6f-03a3-45a5-a5cc-66e1c6177e58.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1346.917628] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02a9df43-3582-463c-a474-ca273f6e27bd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.923946] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1346.923946] env[61991]: value = "task-1130566" [ 1346.923946] env[61991]: _type = "Task" [ 1346.923946] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.931633] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130566, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.433869] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130566, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464768} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.434295] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/254d700f-2f5a-49a3-8762-cec07162124a/254d700f-2f5a-49a3-8762-cec07162124a.vmdk to [datastore2] 9b378b6f-03a3-45a5-a5cc-66e1c6177e58/9b378b6f-03a3-45a5-a5cc-66e1c6177e58.vmdk {{(pid=61991) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1347.434378] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Extending root virtual disk to 1048576 {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1347.434603] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9da5cba-e6d9-40ef-934e-49beaed06eaf {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.441113] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1347.441113] env[61991]: value = "task-1130567" [ 1347.441113] env[61991]: _type = "Task" [ 1347.441113] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.447642] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130567, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.950931] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130567, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065303} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.951229] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Extended root virtual disk {{(pid=61991) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1347.951956] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e857a27-15fd-491f-bdf5-5049f715805c {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.972531] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 9b378b6f-03a3-45a5-a5cc-66e1c6177e58/9b378b6f-03a3-45a5-a5cc-66e1c6177e58.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1347.972722] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f44c0bc-c935-4918-b423-67d12852ec5a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.992296] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1347.992296] env[61991]: value = "task-1130568" [ 1347.992296] env[61991]: _type = "Task" [ 1347.992296] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.999351] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130568, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.504123] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130568, 'name': ReconfigVM_Task, 'duration_secs': 0.26792} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.504508] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 9b378b6f-03a3-45a5-a5cc-66e1c6177e58/9b378b6f-03a3-45a5-a5cc-66e1c6177e58.vmdk or device None with type sparse {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1348.505053] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68c15490-b54e-49c8-b4f0-3f907c3d9db9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.511667] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1348.511667] env[61991]: value = "task-1130569" [ 1348.511667] env[61991]: _type = "Task" [ 1348.511667] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.518747] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130569, 'name': Rename_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.021486] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130569, 'name': Rename_Task, 'duration_secs': 0.128678} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.021884] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Powering on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1349.022089] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0560b12d-8209-47e6-8830-077a09267ec4 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.027989] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1349.027989] env[61991]: value = "task-1130570" [ 1349.027989] env[61991]: _type = "Task" [ 1349.027989] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.034977] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.538219] env[61991]: DEBUG oslo_vmware.api [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130570, 'name': PowerOnVM_Task, 'duration_secs': 0.424932} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.538595] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Powered on the VM {{(pid=61991) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1349.538719] env[61991]: INFO nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1349.538846] env[61991]: DEBUG nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Checking state {{(pid=61991) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1349.539616] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a8418e-4651-4728-b4a0-2939735b85d5 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.058183] env[61991]: INFO nova.compute.manager [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Took 11.30 seconds to build instance. [ 1350.560866] env[61991]: DEBUG oslo_concurrency.lockutils [None req-480f013b-d274-42bc-bd41-00e275017347 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.811s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1350.784860] env[61991]: DEBUG nova.compute.manager [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Received event network-changed-2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1350.785076] env[61991]: DEBUG nova.compute.manager [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Refreshing instance network info cache due to event network-changed-2cfbc2b4-6db8-4951-aae1-61fd8804b599. {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1350.785369] env[61991]: DEBUG oslo_concurrency.lockutils [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] Acquiring lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1350.785467] env[61991]: DEBUG oslo_concurrency.lockutils [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] Acquired lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1350.785670] env[61991]: DEBUG nova.network.neutron [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Refreshing network info cache for port 2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1351.507400] env[61991]: DEBUG nova.network.neutron [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updated VIF entry in instance network info cache for port 2cfbc2b4-6db8-4951-aae1-61fd8804b599. {{(pid=61991) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1351.507788] env[61991]: DEBUG nova.network.neutron [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [{"id": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "address": "fa:16:3e:5d:ab:b9", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cfbc2b4-6d", "ovs_interfaceid": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1352.011032] env[61991]: DEBUG oslo_concurrency.lockutils [req-bccf65f1-702c-4781-b635-c25c7e23ae53 req-4ce4dd46-2b05-449c-a8db-d4267b7af179 service nova] Releasing lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1355.969554] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1355.970042] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1355.970042] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1355.970131] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Cleaning up deleted instances {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1356.476819] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] There are 16 instances to clean {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1356.477090] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 60e3238b-a0ef-4ccf-bce1-633badd10e34] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1356.980254] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: df3b0fba-50ac-4a73-a0e4-f1cfe2705379] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1357.483548] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 6a90ba3f-b947-4f30-8699-29537b97a214] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1357.987199] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: af7962c8-7700-4ff2-988e-65d2922fd514] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1358.491231] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f2c347a3-47ef-48be-8aa6-1b45bc09e7c7] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1358.994611] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: f42c1c55-3635-460e-a106-e63e57426fd2] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1359.497914] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 2d4a14e9-0ea7-42bf-b35f-ce8d374c489d] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1360.001214] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 891eb9f3-d68a-4d10-996b-f2a780a4d3f9] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1360.505473] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: a0901469-3a2a-4d91-90fc-1909c0573caf] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1361.008994] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: ea9e4210-9a4a-40db-9cc4-a59e3306019b] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1361.512221] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 22f59842-073b-4bca-bf4f-a83552a90582] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1362.015798] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 84a0dee7-c33b-494a-ad38-83da6ab44ce2] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1362.519773] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: b26f02c6-a8ea-402d-922d-7879ccb28b92] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1363.022914] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: c43df224-7c35-4b82-ba13-50f3a6f93f2e] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1363.527056] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 100d5fc0-e9d9-4892-9f60-99d657c533ed] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1364.030103] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: cf99c9cc-24c3-4acc-8120-49c4b12a3553] Instance has had 0 of 5 cleanup attempts {{(pid=61991) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1365.533928] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1365.534330] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1365.534330] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1365.534452] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1366.063941] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1366.064115] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1366.064292] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1366.064467] env[61991]: DEBUG nova.objects.instance [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lazy-loading 'info_cache' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1367.767874] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [{"id": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "address": "fa:16:3e:5d:ab:b9", "network": {"id": "a21fe268-1d82-4af5-8409-826ad70120e0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1228634389-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.221", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49893e8554324098b659ce6c24acd5d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cfbc2b4-6d", "ovs_interfaceid": "2cfbc2b4-6db8-4951-aae1-61fd8804b599", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1368.270046] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1368.270313] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1368.270470] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.270632] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.270780] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.270934] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.271073] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1368.271216] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.773843] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1368.774218] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1368.774218] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.774359] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1368.775647] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1645e8a1-2bd9-4397-9d01-47b93d8f06b7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.783456] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9a5ee9-09d2-4563-b1b0-291db4c52bdc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.796703] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e95d4b-836f-431f-994e-30968536cd53 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.802621] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e9e17e-7fd1-4fdc-8050-5f01f6a8bab0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.831242] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181380MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1368.831391] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1368.831556] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1369.919140] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Instance 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61991) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1369.919399] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1369.919531] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1369.945230] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f11df6-840e-4fb6-ab47-6767b9cef518 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.952582] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0b7177-de82-46d0-9ff3-4b9afefa3db9 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.980846] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1349c702-dff8-4f9c-b312-80ff91283780 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.987354] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a6e93a-e6df-423e-acbf-5fb23ebf138e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.000195] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1370.503851] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1371.008511] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1371.008959] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.177s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.008959] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1372.472054] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1372.472478] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Cleaning up deleted instances with incomplete migration {{(pid=61991) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1389.478087] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1389.478398] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1389.982350] env[61991]: DEBUG nova.compute.utils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1390.485405] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1391.548539] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1391.548942] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1391.549118] env[61991]: INFO nova.compute.manager [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Attaching volume 6720520f-0d7f-459a-bc88-cc200c895081 to /dev/sdb [ 1391.579364] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf516d06-fdb7-40f5-bbc9-6d74191fab49 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.586322] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c171d7d3-33fc-4954-949b-99b2cedae5d1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.598957] env[61991]: DEBUG nova.virt.block_device [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating existing volume attachment record: ab7b2001-3416-46b6-8504-4a6a2ddecb65 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1396.141841] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1396.142165] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246997', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'name': 'volume-6720520f-0d7f-459a-bc88-cc200c895081', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'serial': '6720520f-0d7f-459a-bc88-cc200c895081'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1396.143067] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af15b29a-e8df-4295-a24b-63a0947265c2 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.160127] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b9deab-a208-406e-9ed0-0f23a138d844 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.183792] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] volume-6720520f-0d7f-459a-bc88-cc200c895081/volume-6720520f-0d7f-459a-bc88-cc200c895081.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1396.184041] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de3d68e0-6af6-4149-b8ad-8ea76d2ee660 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.201808] env[61991]: DEBUG oslo_vmware.api [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1396.201808] env[61991]: value = "task-1130573" [ 1396.201808] env[61991]: _type = "Task" [ 1396.201808] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.209475] env[61991]: DEBUG oslo_vmware.api [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.713045] env[61991]: DEBUG oslo_vmware.api [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130573, 'name': ReconfigVM_Task, 'duration_secs': 0.331498} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.713045] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfigured VM instance instance-00000077 to attach disk [datastore1] volume-6720520f-0d7f-459a-bc88-cc200c895081/volume-6720520f-0d7f-459a-bc88-cc200c895081.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1396.717165] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d20352c1-212a-4899-975b-fcade401ebaa {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.731315] env[61991]: DEBUG oslo_vmware.api [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1396.731315] env[61991]: value = "task-1130574" [ 1396.731315] env[61991]: _type = "Task" [ 1396.731315] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.738467] env[61991]: DEBUG oslo_vmware.api [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.241015] env[61991]: DEBUG oslo_vmware.api [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130574, 'name': ReconfigVM_Task, 'duration_secs': 0.124627} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.241321] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246997', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'name': 'volume-6720520f-0d7f-459a-bc88-cc200c895081', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'serial': '6720520f-0d7f-459a-bc88-cc200c895081'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1398.277221] env[61991]: DEBUG nova.objects.instance [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1398.784188] env[61991]: DEBUG oslo_concurrency.lockutils [None req-9812437f-4c4d-452c-ac61-88d78790424c tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1399.654646] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1399.654935] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1400.158813] env[61991]: DEBUG nova.compute.utils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Using /dev/sd instead of None {{(pid=61991) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1400.662301] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1401.722062] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1401.722062] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1401.722062] env[61991]: INFO nova.compute.manager [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Attaching volume 565c6d56-b0cc-4e9e-bc77-820f12aba8b8 to /dev/sdc [ 1401.775929] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d97ff92-6f46-4fac-809c-8953ca85ed46 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.783144] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc45572-ba95-4072-8510-6d58faae2565 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.796680] env[61991]: DEBUG nova.virt.block_device [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating existing volume attachment record: e35d0344-b9ac-4c82-87f3-f776d7becb16 {{(pid=61991) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1406.339710] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Volume attach. Driver type: vmdk {{(pid=61991) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1406.339988] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246998', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'name': 'volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'serial': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1406.340891] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b33b95-f56b-441b-b9ba-7790113b8990 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.356838] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b856e59-5d64-432e-a3b0-e70fb555be35 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.382545] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8/volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1406.382774] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f676b787-d09f-4abf-8255-e29fa51fe330 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.399436] env[61991]: DEBUG oslo_vmware.api [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1406.399436] env[61991]: value = "task-1130577" [ 1406.399436] env[61991]: _type = "Task" [ 1406.399436] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.406691] env[61991]: DEBUG oslo_vmware.api [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130577, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.909014] env[61991]: DEBUG oslo_vmware.api [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130577, 'name': ReconfigVM_Task, 'duration_secs': 0.336513} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.909330] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfigured VM instance instance-00000077 to attach disk [datastore1] volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8/volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8.vmdk or device None with type thin {{(pid=61991) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1406.913937] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de498249-f2d6-4474-ae9d-d5999e4cef65 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.928424] env[61991]: DEBUG oslo_vmware.api [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1406.928424] env[61991]: value = "task-1130578" [ 1406.928424] env[61991]: _type = "Task" [ 1406.928424] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.937858] env[61991]: DEBUG oslo_vmware.api [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130578, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.438142] env[61991]: DEBUG oslo_vmware.api [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130578, 'name': ReconfigVM_Task, 'duration_secs': 0.128579} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.438441] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246998', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'name': 'volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'serial': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8'} {{(pid=61991) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1408.474925] env[61991]: DEBUG nova.objects.instance [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1408.981170] env[61991]: DEBUG oslo_concurrency.lockutils [None req-b388d256-edbb-4acf-9a73-95ed55cbcaa4 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1409.259849] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1409.260096] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1409.763635] env[61991]: INFO nova.compute.manager [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Detaching volume 6720520f-0d7f-459a-bc88-cc200c895081 [ 1409.793306] env[61991]: INFO nova.virt.block_device [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Attempting to driver detach volume 6720520f-0d7f-459a-bc88-cc200c895081 from mountpoint /dev/sdb [ 1409.793569] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1409.793759] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246997', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'name': 'volume-6720520f-0d7f-459a-bc88-cc200c895081', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'serial': '6720520f-0d7f-459a-bc88-cc200c895081'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1409.794627] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f900a05b-cb73-4d0e-b7d2-0145aeacb020 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.818973] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41be3777-e0e7-4da3-8648-b29e0ff1cb20 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.825955] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abee03bb-7865-43a6-8585-6d6b01b7f6a0 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.848076] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfef126-1365-4189-a0cd-293810d19900 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.862967] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] The volume has not been displaced from its original location: [datastore1] volume-6720520f-0d7f-459a-bc88-cc200c895081/volume-6720520f-0d7f-459a-bc88-cc200c895081.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1409.867999] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfiguring VM instance instance-00000077 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1409.868261] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7829c061-afbd-46af-b968-7da2f6612bc7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.884753] env[61991]: DEBUG oslo_vmware.api [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1409.884753] env[61991]: value = "task-1130579" [ 1409.884753] env[61991]: _type = "Task" [ 1409.884753] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1409.892132] env[61991]: DEBUG oslo_vmware.api [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.394562] env[61991]: DEBUG oslo_vmware.api [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130579, 'name': ReconfigVM_Task, 'duration_secs': 0.257242} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.394828] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfigured VM instance instance-00000077 to detach disk 2001 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1410.399403] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e881cf6-5257-452d-9e00-df2889a4e521 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.413578] env[61991]: DEBUG oslo_vmware.api [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1410.413578] env[61991]: value = "task-1130580" [ 1410.413578] env[61991]: _type = "Task" [ 1410.413578] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.420791] env[61991]: DEBUG oslo_vmware.api [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130580, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.922632] env[61991]: DEBUG oslo_vmware.api [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130580, 'name': ReconfigVM_Task, 'duration_secs': 0.13239} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.922953] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246997', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'name': 'volume-6720520f-0d7f-459a-bc88-cc200c895081', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '6720520f-0d7f-459a-bc88-cc200c895081', 'serial': '6720520f-0d7f-459a-bc88-cc200c895081'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1411.462011] env[61991]: DEBUG nova.objects.instance [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1412.468814] env[61991]: DEBUG oslo_concurrency.lockutils [None req-bc214359-aa45-4105-b50a-39fdd5bf669a tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.208s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1412.496922] env[61991]: DEBUG oslo_concurrency.lockutils [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1412.497301] env[61991]: DEBUG oslo_concurrency.lockutils [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1413.001228] env[61991]: INFO nova.compute.manager [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Detaching volume 565c6d56-b0cc-4e9e-bc77-820f12aba8b8 [ 1413.030773] env[61991]: INFO nova.virt.block_device [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Attempting to driver detach volume 565c6d56-b0cc-4e9e-bc77-820f12aba8b8 from mountpoint /dev/sdc [ 1413.031055] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Volume detach. Driver type: vmdk {{(pid=61991) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1413.031261] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246998', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'name': 'volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'serial': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1413.032108] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c13114e8-85dd-46c0-bf8b-74ca694c45f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.052692] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedc3669-3670-49b5-8efe-095423df77bc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.060065] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba101d5-2677-490b-a31d-243ac43af840 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.724885] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d1bf41-3f1e-41ed-88be-85972eaea8f8 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.739750] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] The volume has not been displaced from its original location: [datastore1] volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8/volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8.vmdk. No consolidation needed. {{(pid=61991) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1413.744743] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfiguring VM instance instance-00000077 to detach disk 2002 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1413.745027] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f47e7aa6-1753-4d6e-bc44-cd61658ff7c1 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.762537] env[61991]: DEBUG oslo_vmware.api [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1413.762537] env[61991]: value = "task-1130581" [ 1413.762537] env[61991]: _type = "Task" [ 1413.762537] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1413.769570] env[61991]: DEBUG oslo_vmware.api [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.272731] env[61991]: DEBUG oslo_vmware.api [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130581, 'name': ReconfigVM_Task, 'duration_secs': 0.213367} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.273042] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Reconfigured VM instance instance-00000077 to detach disk 2002 {{(pid=61991) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1414.277417] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6fc708a-8edf-400e-87de-32642a1a8b1b {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.291666] env[61991]: DEBUG oslo_vmware.api [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1414.291666] env[61991]: value = "task-1130582" [ 1414.291666] env[61991]: _type = "Task" [ 1414.291666] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.298869] env[61991]: DEBUG oslo_vmware.api [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130582, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.801047] env[61991]: DEBUG oslo_vmware.api [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130582, 'name': ReconfigVM_Task, 'duration_secs': 0.132971} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.801476] env[61991]: DEBUG nova.virt.vmwareapi.volumeops [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-246998', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'name': 'volume-565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9b378b6f-03a3-45a5-a5cc-66e1c6177e58', 'attached_at': '', 'detached_at': '', 'volume_id': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8', 'serial': '565c6d56-b0cc-4e9e-bc77-820f12aba8b8'} {{(pid=61991) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1415.343416] env[61991]: DEBUG nova.objects.instance [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'flavor' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1416.350950] env[61991]: DEBUG oslo_concurrency.lockutils [None req-71ea54c0-5598-4c72-a27b-c62adacd2ed0 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.853s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1416.473045] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1416.869933] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1416.870255] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1416.870480] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1416.870668] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1416.870845] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1416.873084] env[61991]: INFO nova.compute.manager [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Terminating instance [ 1416.874800] env[61991]: DEBUG nova.compute.manager [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Start destroying the instance on the hypervisor. {{(pid=61991) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1416.875008] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Destroying instance {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1416.875827] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21baba46-1bc7-461c-8395-8b102c0f1c7a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.883624] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Powering off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1416.883840] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7f8de97-d475-449f-9ee7-dc619a72f80a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.889872] env[61991]: DEBUG oslo_vmware.api [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1416.889872] env[61991]: value = "task-1130583" [ 1416.889872] env[61991]: _type = "Task" [ 1416.889872] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.897978] env[61991]: DEBUG oslo_vmware.api [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.969799] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1417.399783] env[61991]: DEBUG oslo_vmware.api [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130583, 'name': PowerOffVM_Task, 'duration_secs': 0.169134} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.400105] env[61991]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Powered off the VM {{(pid=61991) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1417.400282] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Unregistering the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1417.400540] env[61991]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-332f45fd-cf23-484c-97da-0a7b6ee901dc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.462426] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Unregistered the VM {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1417.462698] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Deleting contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1417.462927] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Deleting the datastore file [datastore2] 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1417.463256] env[61991]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-070b6b9e-8166-4c7c-9101-8ccad5d95362 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.469791] env[61991]: DEBUG oslo_vmware.api [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for the task: (returnval){ [ 1417.469791] env[61991]: value = "task-1130585" [ 1417.469791] env[61991]: _type = "Task" [ 1417.469791] env[61991]: } to complete. {{(pid=61991) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.477588] env[61991]: DEBUG oslo_vmware.api [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.980940] env[61991]: DEBUG oslo_vmware.api [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Task: {'id': task-1130585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127952} completed successfully. {{(pid=61991) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1417.980940] env[61991]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Deleted the datastore file {{(pid=61991) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1417.980940] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Deleted contents of the VM from datastore datastore2 {{(pid=61991) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1417.980940] env[61991]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Instance destroyed {{(pid=61991) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1417.981284] env[61991]: INFO nova.compute.manager [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1417.981284] env[61991]: DEBUG oslo.service.loopingcall [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61991) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1417.981469] env[61991]: DEBUG nova.compute.manager [-] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Deallocating network for instance {{(pid=61991) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1417.981559] env[61991]: DEBUG nova.network.neutron [-] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] deallocate_for_instance() {{(pid=61991) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1418.611795] env[61991]: DEBUG nova.compute.manager [req-1f901622-2d52-4de8-82ea-be7f8f594da9 req-ea4aaea1-4f4e-4fe8-b2bf-5aff555f3837 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Received event network-vif-deleted-2cfbc2b4-6db8-4951-aae1-61fd8804b599 {{(pid=61991) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1418.612041] env[61991]: INFO nova.compute.manager [req-1f901622-2d52-4de8-82ea-be7f8f594da9 req-ea4aaea1-4f4e-4fe8-b2bf-5aff555f3837 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Neutron deleted interface 2cfbc2b4-6db8-4951-aae1-61fd8804b599; detaching it from the instance and deleting it from the info cache [ 1418.612114] env[61991]: DEBUG nova.network.neutron [req-1f901622-2d52-4de8-82ea-be7f8f594da9 req-ea4aaea1-4f4e-4fe8-b2bf-5aff555f3837 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1418.969813] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1418.970129] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1418.970349] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1418.970491] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61991) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1419.055327] env[61991]: DEBUG nova.network.neutron [-] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1419.115104] env[61991]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-61964932-f98a-4165-a3f3-bf16b7bb5f3f {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.125413] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833de943-fdc9-4410-bb82-d0c38e44cf12 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.148802] env[61991]: DEBUG nova.compute.manager [req-1f901622-2d52-4de8-82ea-be7f8f594da9 req-ea4aaea1-4f4e-4fe8-b2bf-5aff555f3837 service nova] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Detach interface failed, port_id=2cfbc2b4-6db8-4951-aae1-61fd8804b599, reason: Instance 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 could not be found. {{(pid=61991) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1419.558462] env[61991]: INFO nova.compute.manager [-] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Took 1.58 seconds to deallocate network for instance. [ 1419.965957] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1419.969584] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1419.969734] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Starting heal instance info cache {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1419.969852] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Rebuilding the list of instances to heal {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1420.066337] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1420.066591] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1420.066813] env[61991]: DEBUG nova.objects.instance [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lazy-loading 'resources' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1420.513739] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1420.513896] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquired lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1420.514062] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Forcefully refreshing network info cache for instance {{(pid=61991) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1420.514221] env[61991]: DEBUG nova.objects.instance [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lazy-loading 'info_cache' on Instance uuid 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 {{(pid=61991) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1420.601072] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b8b0c2-03f6-49a7-b9db-c0a32a4f0c96 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.609627] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0de105-e2fc-4ffe-8d9f-46e49e3b59c7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.639363] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6a906f-5aca-4241-8574-680f8e861546 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.645898] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ebc1f3-fe1d-4f5f-be2b-72d8a791f668 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.658373] env[61991]: DEBUG nova.compute.provider_tree [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1421.161596] env[61991]: DEBUG nova.scheduler.client.report [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1421.537148] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Instance cache missing network info. {{(pid=61991) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1421.666332] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1421.688326] env[61991]: INFO nova.scheduler.client.report [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Deleted allocations for instance 9b378b6f-03a3-45a5-a5cc-66e1c6177e58 [ 1422.115237] env[61991]: DEBUG nova.network.neutron [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updating instance_info_cache with network_info: [] {{(pid=61991) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1422.196405] env[61991]: DEBUG oslo_concurrency.lockutils [None req-3b5acdad-7ace-4924-b01e-fb7436ef2863 tempest-AttachVolumeTestJSON-2147355869 tempest-AttachVolumeTestJSON-2147355869-project-member] Lock "9b378b6f-03a3-45a5-a5cc-66e1c6177e58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.326s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1422.618362] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Releasing lock "refresh_cache-9b378b6f-03a3-45a5-a5cc-66e1c6177e58" {{(pid=61991) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1422.618584] env[61991]: DEBUG nova.compute.manager [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] [instance: 9b378b6f-03a3-45a5-a5cc-66e1c6177e58] Updated the network info_cache for instance {{(pid=61991) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1422.618818] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1423.121987] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.122253] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1423.122432] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1423.122588] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61991) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1423.123515] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a17dc7-41e0-4580-a38a-e47865cfba16 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.131737] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af37c56-e7a8-425d-b455-7da54fab6924 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.145021] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5bf2e0-ee34-4c32-b56b-cdfe8f67289a {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.150921] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96185a8e-5477-4b78-b001-436f6a1831fc {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.178084] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181387MB free_disk=168GB free_vcpus=48 pci_devices=None {{(pid=61991) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1423.178239] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1423.178420] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1424.198133] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1424.198378] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61991) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1424.211156] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84ba89b-c803-45a6-b1da-61a60fa4198e {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.218670] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da288997-63a6-4b2a-9188-1329ba2049f7 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.248244] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c2ac16-4221-4ee7-9ceb-28618b7e5e46 {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.254997] env[61991]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3abe21-4966-4d13-9c37-9a1a7e8df9dd {{(pid=61991) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.267446] env[61991]: DEBUG nova.compute.provider_tree [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed in ProviderTree for provider: d748992a-e0bf-4ec2-9c17-0e373360e5a3 {{(pid=61991) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1424.770732] env[61991]: DEBUG nova.scheduler.client.report [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Inventory has not changed for provider d748992a-e0bf-4ec2-9c17-0e373360e5a3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 168, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61991) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1425.275345] env[61991]: DEBUG nova.compute.resource_tracker [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61991) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1425.275724] env[61991]: DEBUG oslo_concurrency.lockutils [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.097s {{(pid=61991) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1427.626510] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1431.965440] env[61991]: DEBUG oslo_service.periodic_task [None req-d40280b5-71c1-4249-b1b4-df4ae041f858 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61991) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}